[GitHub] [airflow] galuszkak commented on issue #6660: [AIRFLOW-6065] Add Stackdriver Task Handler
galuszkak commented on issue #6660: [AIRFLOW-6065] Add Stackdriver Task Handler URL: https://github.com/apache/airflow/pull/6660#issuecomment-588672062 @mik-laj we have every week mini hackathon so if this is not urgent we are doing this every Thursday 8:00-9:30 with 5-6 people. Now we are working on enabling multiple remote loggers in Airflow, and we can also try to take this Stackdriver logs in Web UI. This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[jira] [Assigned] (AIRFLOW-6848) Can i create the viewer filter in LDAP ad group
[ https://issues.apache.org/jira/browse/AIRFLOW-6848?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Deepak Parashar reassigned AIRFLOW-6848: Assignee: Kamil Bregula > Can i create the viewer filter in LDAP ad group > > > Key: AIRFLOW-6848 > URL: https://issues.apache.org/jira/browse/AIRFLOW-6848 > Project: Apache Airflow > Issue Type: New Feature > Components: authentication >Affects Versions: 1.10.5, 1.10.8 >Reporter: Deepak Parashar >Assignee: Kamil Bregula >Priority: Major > Labels: authentication, authorization, ldap > > Hi , > I am using ad based authentication to access the airflow in my company , i > have created two groups in ad(ldap) for superuser_filter and data_profiler , > i want to create or integrate one group like viewer which can help me to > allow only read only access of dags and other system information for the > users who are part of this viewer group . > has someone already did it ? > group_member_attr = memberOf superuser_filter = > memberOf=CN=airflow-super-users,OU=Groups,OU=RWC,OU=US,OU=NORAM,DC=example,DC=com > data_profiler_filter = > memberOf=CN=airflow-data-profilers,OU=Groups,OU=RWC,OU=US,OU=NORAM,DC=example,DC=com > > like these two above groups , similarly we can create one for viewer . > if its not already available can someone please help me how can i implement > it in the current setup . > -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Updated] (AIRFLOW-6848) Can i create the viewer filter in LDAP ad group
[ https://issues.apache.org/jira/browse/AIRFLOW-6848?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Deepak Parashar updated AIRFLOW-6848: - Labels: authentication authorization ldap (was: ) > Can i create the viewer filter in LDAP ad group > > > Key: AIRFLOW-6848 > URL: https://issues.apache.org/jira/browse/AIRFLOW-6848 > Project: Apache Airflow > Issue Type: New Feature > Components: authentication >Affects Versions: 1.10.5, 1.10.8 >Reporter: Deepak Parashar >Priority: Major > Labels: authentication, authorization, ldap > > Hi , > I am using ad based authentication to access the airflow in my company , i > have created two groups in ad(ldap) for superuser_filter and data_profiler , > i want to create or integrate one group like viewer which can help me to > allow only read only access of dags and other system information for the > users who are part of this viewer group . > has someone already did it ? > group_member_attr = memberOf superuser_filter = > memberOf=CN=airflow-super-users,OU=Groups,OU=RWC,OU=US,OU=NORAM,DC=example,DC=com > data_profiler_filter = > memberOf=CN=airflow-data-profilers,OU=Groups,OU=RWC,OU=US,OU=NORAM,DC=example,DC=com > > like these two above groups , similarly we can create one for viewer . > if its not already available can someone please help me how can i implement > it in the current setup . > -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Created] (AIRFLOW-6848) Can i create the viewer filter in LDAP ad group
Deepak Parashar created AIRFLOW-6848: Summary: Can i create the viewer filter in LDAP ad group Key: AIRFLOW-6848 URL: https://issues.apache.org/jira/browse/AIRFLOW-6848 Project: Apache Airflow Issue Type: New Feature Components: authentication Affects Versions: 1.10.8, 1.10.5 Reporter: Deepak Parashar Hi , I am using ad based authentication to access the airflow in my company , i have created two groups in ad(ldap) for superuser_filter and data_profiler , i want to create or integrate one group like viewer which can help me to allow only read only access of dags and other system information for the users who are part of this viewer group . has someone already did it ? group_member_attr = memberOf superuser_filter = memberOf=CN=airflow-super-users,OU=Groups,OU=RWC,OU=US,OU=NORAM,DC=example,DC=com data_profiler_filter = memberOf=CN=airflow-data-profilers,OU=Groups,OU=RWC,OU=US,OU=NORAM,DC=example,DC=com like these two above groups , similarly we can create one for viewer . if its not already available can someone please help me how can i implement it in the current setup . -- This message was sent by Atlassian Jira (v8.3.4#803005)
[GitHub] [airflow] codecov-io edited a comment on issue #7471: [AIRFLOW-6840] Bump up version of future
codecov-io edited a comment on issue #7471: [AIRFLOW-6840] Bump up version of future URL: https://github.com/apache/airflow/pull/7471#issuecomment-588604085 # [Codecov](https://codecov.io/gh/apache/airflow/pull/7471?src=pr&el=h1) Report > Merging [#7471](https://codecov.io/gh/apache/airflow/pull/7471?src=pr&el=desc) into [v1-10-test](https://codecov.io/gh/apache/airflow/commit/6135f7dcd76db27acdb70e4394db7449fa1bcf7a?src=pr&el=desc) will **increase** coverage by `<.01%`. > The diff coverage is `n/a`. [![Impacted file tree graph](https://codecov.io/gh/apache/airflow/pull/7471/graphs/tree.svg?width=650&token=WdLKlKHOAU&height=150&src=pr)](https://codecov.io/gh/apache/airflow/pull/7471?src=pr&el=tree) ```diff @@ Coverage Diff @@ ## v1-10-test#7471 +/- ## == + Coverage 81.63% 81.64% +<.01% == Files 529 529 Lines 3643736437 == + Hits2974729749 +2 + Misses 6690 6688 -2 ``` | [Impacted Files](https://codecov.io/gh/apache/airflow/pull/7471?src=pr&el=tree) | Coverage Δ | | |---|---|---| | [airflow/operators/postgres\_operator.py](https://codecov.io/gh/apache/airflow/pull/7471/diff?src=pr&el=tree#diff-YWlyZmxvdy9vcGVyYXRvcnMvcG9zdGdyZXNfb3BlcmF0b3IucHk=) | `100% <0%> (ø)` | :arrow_up: | | [airflow/operators/mysql\_operator.py](https://codecov.io/gh/apache/airflow/pull/7471/diff?src=pr&el=tree#diff-YWlyZmxvdy9vcGVyYXRvcnMvbXlzcWxfb3BlcmF0b3IucHk=) | `100% <0%> (ø)` | :arrow_up: | | [airflow/operators/mysql\_to\_hive.py](https://codecov.io/gh/apache/airflow/pull/7471/diff?src=pr&el=tree#diff-YWlyZmxvdy9vcGVyYXRvcnMvbXlzcWxfdG9faGl2ZS5weQ==) | `100% <0%> (ø)` | :arrow_up: | | [airflow/operators/generic\_transfer.py](https://codecov.io/gh/apache/airflow/pull/7471/diff?src=pr&el=tree#diff-YWlyZmxvdy9vcGVyYXRvcnMvZ2VuZXJpY190cmFuc2Zlci5weQ==) | `100% <0%> (ø)` | :arrow_up: | | [airflow/contrib/kubernetes/volume\_mount.py](https://codecov.io/gh/apache/airflow/pull/7471/diff?src=pr&el=tree#diff-YWlyZmxvdy9jb250cmliL2t1YmVybmV0ZXMvdm9sdW1lX21vdW50LnB5) | `100% <0%> (ø)` | :arrow_up: | | [airflow/hooks/hdfs\_hook.py](https://codecov.io/gh/apache/airflow/pull/7471/diff?src=pr&el=tree#diff-YWlyZmxvdy9ob29rcy9oZGZzX2hvb2sucHk=) | `92.5% <0%> (ø)` | :arrow_up: | | [airflow/contrib/kubernetes/volume.py](https://codecov.io/gh/apache/airflow/pull/7471/diff?src=pr&el=tree#diff-YWlyZmxvdy9jb250cmliL2t1YmVybmV0ZXMvdm9sdW1lLnB5) | `100% <0%> (ø)` | :arrow_up: | | [airflow/contrib/kubernetes/pod\_launcher.py](https://codecov.io/gh/apache/airflow/pull/7471/diff?src=pr&el=tree#diff-YWlyZmxvdy9jb250cmliL2t1YmVybmV0ZXMvcG9kX2xhdW5jaGVyLnB5) | `92.48% <0%> (ø)` | :arrow_up: | | [airflow/security/kerberos.py](https://codecov.io/gh/apache/airflow/pull/7471/diff?src=pr&el=tree#diff-YWlyZmxvdy9zZWN1cml0eS9rZXJiZXJvcy5weQ==) | `75.55% <0%> (ø)` | :arrow_up: | | [airflow/contrib/kubernetes/pod\_generator.py](https://codecov.io/gh/apache/airflow/pull/7471/diff?src=pr&el=tree#diff-YWlyZmxvdy9jb250cmliL2t1YmVybmV0ZXMvcG9kX2dlbmVyYXRvci5weQ==) | `87.5% <0%> (ø)` | :arrow_up: | | ... and [25 more](https://codecov.io/gh/apache/airflow/pull/7471/diff?src=pr&el=tree-more) | | -- [Continue to review full report at Codecov](https://codecov.io/gh/apache/airflow/pull/7471?src=pr&el=continue). > **Legend** - [Click here to learn more](https://docs.codecov.io/docs/codecov-delta) > `Δ = absolute (impact)`, `ø = not affected`, `? = missing data` > Powered by [Codecov](https://codecov.io/gh/apache/airflow/pull/7471?src=pr&el=footer). Last update [6135f7d...2506022](https://codecov.io/gh/apache/airflow/pull/7471?src=pr&el=lastupdated). Read the [comment docs](https://docs.codecov.io/docs/pull-request-comments). This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[GitHub] [airflow] codecov-io commented on issue #7471: [AIRFLOW-6840] Bump up version of future
codecov-io commented on issue #7471: [AIRFLOW-6840] Bump up version of future URL: https://github.com/apache/airflow/pull/7471#issuecomment-588604085 # [Codecov](https://codecov.io/gh/apache/airflow/pull/7471?src=pr&el=h1) Report > Merging [#7471](https://codecov.io/gh/apache/airflow/pull/7471?src=pr&el=desc) into [v1-10-test](https://codecov.io/gh/apache/airflow/commit/6135f7dcd76db27acdb70e4394db7449fa1bcf7a?src=pr&el=desc) will **increase** coverage by `<.01%`. > The diff coverage is `n/a`. [![Impacted file tree graph](https://codecov.io/gh/apache/airflow/pull/7471/graphs/tree.svg?width=650&token=WdLKlKHOAU&height=150&src=pr)](https://codecov.io/gh/apache/airflow/pull/7471?src=pr&el=tree) ```diff @@ Coverage Diff @@ ## v1-10-test#7471 +/- ## == + Coverage 81.63% 81.64% +<.01% == Files 529 529 Lines 3643736437 == + Hits2974729749 +2 + Misses 6690 6688 -2 ``` | [Impacted Files](https://codecov.io/gh/apache/airflow/pull/7471?src=pr&el=tree) | Coverage Δ | | |---|---|---| | [airflow/utils/dag\_processing.py](https://codecov.io/gh/apache/airflow/pull/7471/diff?src=pr&el=tree#diff-YWlyZmxvdy91dGlscy9kYWdfcHJvY2Vzc2luZy5weQ==) | `92.33% <0%> (+0.32%)` | :arrow_up: | -- [Continue to review full report at Codecov](https://codecov.io/gh/apache/airflow/pull/7471?src=pr&el=continue). > **Legend** - [Click here to learn more](https://docs.codecov.io/docs/codecov-delta) > `Δ = absolute (impact)`, `ø = not affected`, `? = missing data` > Powered by [Codecov](https://codecov.io/gh/apache/airflow/pull/7471?src=pr&el=footer). Last update [6135f7d...2506022](https://codecov.io/gh/apache/airflow/pull/7471?src=pr&el=lastupdated). Read the [comment docs](https://docs.codecov.io/docs/pull-request-comments). This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[GitHub] [airflow] codecov-io commented on issue #7470: [AIRFLOW-6834] Fix flaky test_scheduler_job by sorting TaskInstance
codecov-io commented on issue #7470: [AIRFLOW-6834] Fix flaky test_scheduler_job by sorting TaskInstance URL: https://github.com/apache/airflow/pull/7470#issuecomment-588602591 # [Codecov](https://codecov.io/gh/apache/airflow/pull/7470?src=pr&el=h1) Report > Merging [#7470](https://codecov.io/gh/apache/airflow/pull/7470?src=pr&el=desc) into [master](https://codecov.io/gh/apache/airflow/commit/1a9a9f7618f1c22e3e9a6ef4ec73b717c7760c7d?src=pr&el=desc) will **decrease** coverage by `0.45%`. > The diff coverage is `100%`. [![Impacted file tree graph](https://codecov.io/gh/apache/airflow/pull/7470/graphs/tree.svg?width=650&token=WdLKlKHOAU&height=150&src=pr)](https://codecov.io/gh/apache/airflow/pull/7470?src=pr&el=tree) ```diff @@Coverage Diff @@ ## master#7470 +/- ## == - Coverage 86.68% 86.22% -0.46% == Files 882 882 Lines 4152641526 == - Hits3599735806 -191 - Misses 5529 5720 +191 ``` | [Impacted Files](https://codecov.io/gh/apache/airflow/pull/7470?src=pr&el=tree) | Coverage Δ | | |---|---|---| | [airflow/models/dagrun.py](https://codecov.io/gh/apache/airflow/pull/7470/diff?src=pr&el=tree#diff-YWlyZmxvdy9tb2RlbHMvZGFncnVuLnB5) | `96.55% <100%> (ø)` | :arrow_up: | | [...w/providers/apache/hive/operators/mysql\_to\_hive.py](https://codecov.io/gh/apache/airflow/pull/7470/diff?src=pr&el=tree#diff-YWlyZmxvdy9wcm92aWRlcnMvYXBhY2hlL2hpdmUvb3BlcmF0b3JzL215c3FsX3RvX2hpdmUucHk=) | `35.84% <0%> (-64.16%)` | :arrow_down: | | [airflow/kubernetes/volume\_mount.py](https://codecov.io/gh/apache/airflow/pull/7470/diff?src=pr&el=tree#diff-YWlyZmxvdy9rdWJlcm5ldGVzL3ZvbHVtZV9tb3VudC5weQ==) | `44.44% <0%> (-55.56%)` | :arrow_down: | | [airflow/kubernetes/volume.py](https://codecov.io/gh/apache/airflow/pull/7470/diff?src=pr&el=tree#diff-YWlyZmxvdy9rdWJlcm5ldGVzL3ZvbHVtZS5weQ==) | `52.94% <0%> (-47.06%)` | :arrow_down: | | [airflow/security/kerberos.py](https://codecov.io/gh/apache/airflow/pull/7470/diff?src=pr&el=tree#diff-YWlyZmxvdy9zZWN1cml0eS9rZXJiZXJvcy5weQ==) | `30.43% <0%> (-45.66%)` | :arrow_down: | | [airflow/kubernetes/pod\_launcher.py](https://codecov.io/gh/apache/airflow/pull/7470/diff?src=pr&el=tree#diff-YWlyZmxvdy9rdWJlcm5ldGVzL3BvZF9sYXVuY2hlci5weQ==) | `47.18% <0%> (-45.08%)` | :arrow_down: | | [airflow/providers/mysql/operators/mysql.py](https://codecov.io/gh/apache/airflow/pull/7470/diff?src=pr&el=tree#diff-YWlyZmxvdy9wcm92aWRlcnMvbXlzcWwvb3BlcmF0b3JzL215c3FsLnB5) | `55% <0%> (-45%)` | :arrow_down: | | [...viders/cncf/kubernetes/operators/kubernetes\_pod.py](https://codecov.io/gh/apache/airflow/pull/7470/diff?src=pr&el=tree#diff-YWlyZmxvdy9wcm92aWRlcnMvY25jZi9rdWJlcm5ldGVzL29wZXJhdG9ycy9rdWJlcm5ldGVzX3BvZC5weQ==) | `69.38% <0%> (-25.52%)` | :arrow_down: | | [airflow/kubernetes/refresh\_config.py](https://codecov.io/gh/apache/airflow/pull/7470/diff?src=pr&el=tree#diff-YWlyZmxvdy9rdWJlcm5ldGVzL3JlZnJlc2hfY29uZmlnLnB5) | `50.98% <0%> (-23.53%)` | :arrow_down: | | [airflow/providers/apache/hive/hooks/hive.py](https://codecov.io/gh/apache/airflow/pull/7470/diff?src=pr&el=tree#diff-YWlyZmxvdy9wcm92aWRlcnMvYXBhY2hlL2hpdmUvaG9va3MvaGl2ZS5weQ==) | `76.02% <0%> (-1.54%)` | :arrow_down: | | ... and [2 more](https://codecov.io/gh/apache/airflow/pull/7470/diff?src=pr&el=tree-more) | | -- [Continue to review full report at Codecov](https://codecov.io/gh/apache/airflow/pull/7470?src=pr&el=continue). > **Legend** - [Click here to learn more](https://docs.codecov.io/docs/codecov-delta) > `Δ = absolute (impact)`, `ø = not affected`, `? = missing data` > Powered by [Codecov](https://codecov.io/gh/apache/airflow/pull/7470?src=pr&el=footer). Last update [1a9a9f7...13e180b](https://codecov.io/gh/apache/airflow/pull/7470?src=pr&el=lastupdated). Read the [comment docs](https://docs.codecov.io/docs/pull-request-comments). This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[GitHub] [airflow] yuqian90 opened a new pull request #7471: [AIRFLOW-6840] Bump up version of future
yuqian90 opened a new pull request #7471: [AIRFLOW-6840] Bump up version of future URL: https://github.com/apache/airflow/pull/7471 --- Issue link: WILL BE INSERTED BY [boring-cyborg](https://github.com/kaxil/boring-cyborg) Make sure to mark the boxes below before creating PR: [x] - [ ] Description above provides context of the change - [ ] Commit message/PR title starts with `[AIRFLOW-]`. AIRFLOW- = JIRA ID* - [ ] Unit tests coverage for changes (not needed for documentation changes) - [ ] Commits follow "[How to write a good git commit message](http://chris.beams.io/posts/git-commit/)" - [ ] Relevant documentation is updated including usage instructions. - [ ] I will engage committers as explained in [Contribution Workflow Example](https://github.com/apache/airflow/blob/master/CONTRIBUTING.rst#contribution-workflow-example). * For document-only changes commit message can start with `[AIRFLOW-]`. --- In case of fundamental code change, Airflow Improvement Proposal ([AIP](https://cwiki.apache.org/confluence/display/AIRFLOW/Airflow+Improvements+Proposals)) is needed. In case of a new dependency, check compliance with the [ASF 3rd Party License Policy](https://www.apache.org/legal/resolved.html#category-x). In case of backwards incompatible changes please leave a note in [UPDATING.md](https://github.com/apache/airflow/blob/master/UPDATING.md). Read the [Pull Request Guidelines](https://github.com/apache/airflow/blob/master/CONTRIBUTING.rst#pull-request-guidelines) for more information. This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[jira] [Commented] (AIRFLOW-6840) Bump up version of future to 0.18.2 in v-10-testing to get closer to python 3.8
[ https://issues.apache.org/jira/browse/AIRFLOW-6840?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17040608#comment-17040608 ] ASF GitHub Bot commented on AIRFLOW-6840: - yuqian90 commented on pull request #7471: [AIRFLOW-6840] Bump up version of future URL: https://github.com/apache/airflow/pull/7471 --- Issue link: WILL BE INSERTED BY [boring-cyborg](https://github.com/kaxil/boring-cyborg) Make sure to mark the boxes below before creating PR: [x] - [ ] Description above provides context of the change - [ ] Commit message/PR title starts with `[AIRFLOW-]`. AIRFLOW- = JIRA ID* - [ ] Unit tests coverage for changes (not needed for documentation changes) - [ ] Commits follow "[How to write a good git commit message](http://chris.beams.io/posts/git-commit/)" - [ ] Relevant documentation is updated including usage instructions. - [ ] I will engage committers as explained in [Contribution Workflow Example](https://github.com/apache/airflow/blob/master/CONTRIBUTING.rst#contribution-workflow-example). * For document-only changes commit message can start with `[AIRFLOW-]`. --- In case of fundamental code change, Airflow Improvement Proposal ([AIP](https://cwiki.apache.org/confluence/display/AIRFLOW/Airflow+Improvements+Proposals)) is needed. In case of a new dependency, check compliance with the [ASF 3rd Party License Policy](https://www.apache.org/legal/resolved.html#category-x). In case of backwards incompatible changes please leave a note in [UPDATING.md](https://github.com/apache/airflow/blob/master/UPDATING.md). Read the [Pull Request Guidelines](https://github.com/apache/airflow/blob/master/CONTRIBUTING.rst#pull-request-guidelines) for more information. This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org > Bump up version of future to 0.18.2 in v-10-testing to get closer to python > 3.8 > --- > > Key: AIRFLOW-6840 > URL: https://issues.apache.org/jira/browse/AIRFLOW-6840 > Project: Apache Airflow > Issue Type: Bug > Components: core >Affects Versions: 1.10.9 >Reporter: Qian Yu >Assignee: Qian Yu >Priority: Major > > Bump up future from 0.17 to 0.18.2. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Assigned] (AIRFLOW-6834) Fix flaky test test_scheduler_job.py::TestDagFileProcessor::test_dag_file_processor_process_task_instances_depends_on_past
[ https://issues.apache.org/jira/browse/AIRFLOW-6834?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Qian Yu reassigned AIRFLOW-6834: Assignee: Qian Yu > Fix flaky test > test_scheduler_job.py::TestDagFileProcessor::test_dag_file_processor_process_task_instances_depends_on_past > -- > > Key: AIRFLOW-6834 > URL: https://issues.apache.org/jira/browse/AIRFLOW-6834 > Project: Apache Airflow > Issue Type: Bug > Components: tests >Affects Versions: 1.10.9 >Reporter: Qian Yu >Assignee: Qian Yu >Priority: Major > > test_scheduler_job.py has a few flaky tests. Some are marked with > pytest.mark.xfail, but this one is not marked flaky. It sometimes fails in > Travis. For example: > > {code:python} > FAILURES > > _ > TestDagFileProcessor.test_dag_file_processor_process_task_instances_depends_on_past_0 > __ > a = ( testMethod=test_dag_file_processor_process_task_instances_depends_on_past_0>,) > @wraps(func) > def standalone_func(*a): > > return func(*(a + p.args), **p.kwargs) > /usr/local/lib/python3.6/site-packages/parameterized/parameterized.py:518: > _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ > _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ > self = testMethod=test_dag_file_processor_process_task_instances_depends_on_past_0> > state = None, start_date = None, end_date = None > @parameterized.expand([ > [State.NONE, None, None], > [State.UP_FOR_RETRY, timezone.utcnow() - > datetime.timedelta(minutes=30), > timezone.utcnow() - datetime.timedelta(minutes=15)], > [State.UP_FOR_RESCHEDULE, timezone.utcnow() - > datetime.timedelta(minutes=30), > timezone.utcnow() - datetime.timedelta(minutes=15)], > ]) > def test_dag_file_processor_process_task_instances_depends_on_past(self, > state, start_date, end_date): > """ > Test if _process_task_instances puts the right task instances into the > mock_list. > """ > dag = DAG( > dag_id='test_scheduler_process_execute_task_depends_on_past', > start_date=DEFAULT_DATE, > default_args={ > 'depends_on_past': True, > }, > ) > dag_task1 = DummyOperator( > task_id='dummy1', > dag=dag, > owner='airflow') > dag_task2 = DummyOperator( > task_id='dummy2', > dag=dag, > owner='airflow') > with create_session() as session: > orm_dag = DagModel(dag_id=dag.dag_id) > session.merge(orm_dag) > dag_file_processor = DagFileProcessor(dag_ids=[], > log=mock.MagicMock()) > dag.clear() > dr = dag_file_processor.create_dag_run(dag) > self.assertIsNotNone(dr) > with create_session() as session: > tis = dr.get_task_instances(session=session) > for ti in tis: > ti.state = state > ti.start_date = start_date > ti.end_date = end_date > ti_to_schedule = [] > dag_file_processor._process_task_instances(dag, > task_instances_list=ti_to_schedule) > > assert ti_to_schedule == [ > (dag.dag_id, dag_task1.task_id, DEFAULT_DATE, TRY_NUMBER), > (dag.dag_id, dag_task2.task_id, DEFAULT_DATE, TRY_NUMBER), > ] > E AssertionError: assert > [('test_scheduler_process_execute_task_depends_on_past',\n 'dummy2',\n > datetime.datetime(2016, 1, 1, 0, 0, tzinfo= +00:00:00, STD]>),\n 1),\n > ('test_scheduler_process_execute_task_depends_on_past',\n 'dummy1',\n > datetime.datetime(2016, 1, 1, 0, 0, tzinfo= +00:00:00, STD]>),\n 1)] == > [('test_scheduler_process_execute_task_depends_on_past',\n 'dummy1',\n > datetime.datetime(2016, 1, 1, 0, 0, tzinfo=),\n 1),\n > ('test_scheduler_process_execute_task_depends_on_past',\n 'dummy2',\n > datetime.datetime(2016, 1, 1, 0, 0, tzinfo=),\n 1)] > E At index 0 diff: > ('test_scheduler_process_execute_task_depends_on_past', 'dummy2', > datetime.datetime(2016, 1, 1, 0, 0, tzinfo= +00:00:00, STD]>), 1) != > ('test_scheduler_process_execute_task_depends_on_past', 'dummy1', > datetime.datetime(2016, 1, 1, 0, 0, tzinfo=), 1) > E Full diff: > E [ > E('test_scheduler_process_execute_task_depends_on_past', > E - 'dummy2', > E ? ^ > E + 'dummy1', > E ? ^ > E - datetime.datetime(2016, 1, 1, 0, 0, tzinfo= GMT, +00:00:00, STD]>), >
[jira] [Commented] (AIRFLOW-6834) Fix flaky test test_scheduler_job.py::TestDagFileProcessor::test_dag_file_processor_process_task_instances_depends_on_past
[ https://issues.apache.org/jira/browse/AIRFLOW-6834?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17040604#comment-17040604 ] ASF GitHub Bot commented on AIRFLOW-6834: - yuqian90 commented on pull request #7470: [AIRFLOW-6834] Fix flaky test_scheduler_job by sorting TaskInstance URL: https://github.com/apache/airflow/pull/7470 Fix a flaky test that fails because `TaskInstance` are returned in non-deterministic order: ``` test_scheduler_job.py::TestDagFileProcessor::test_dag_file_processor_process_task_instances_depends_on_past ``` --- Issue link: WILL BE INSERTED BY [boring-cyborg](https://github.com/kaxil/boring-cyborg) Make sure to mark the boxes below before creating PR: [x] - [x] Description above provides context of the change - [x] Commit message/PR title starts with `[AIRFLOW-]`. AIRFLOW- = JIRA ID* - [x] Unit tests coverage for changes (not needed for documentation changes) - [x] Commits follow "[How to write a good git commit message](http://chris.beams.io/posts/git-commit/)" - [x] Relevant documentation is updated including usage instructions. - [x] I will engage committers as explained in [Contribution Workflow Example](https://github.com/apache/airflow/blob/master/CONTRIBUTING.rst#contribution-workflow-example). * For document-only changes commit message can start with `[AIRFLOW-]`. --- In case of fundamental code change, Airflow Improvement Proposal ([AIP](https://cwiki.apache.org/confluence/display/AIRFLOW/Airflow+Improvements+Proposals)) is needed. In case of a new dependency, check compliance with the [ASF 3rd Party License Policy](https://www.apache.org/legal/resolved.html#category-x). In case of backwards incompatible changes please leave a note in [UPDATING.md](https://github.com/apache/airflow/blob/master/UPDATING.md). Read the [Pull Request Guidelines](https://github.com/apache/airflow/blob/master/CONTRIBUTING.rst#pull-request-guidelines) for more information. This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org > Fix flaky test > test_scheduler_job.py::TestDagFileProcessor::test_dag_file_processor_process_task_instances_depends_on_past > -- > > Key: AIRFLOW-6834 > URL: https://issues.apache.org/jira/browse/AIRFLOW-6834 > Project: Apache Airflow > Issue Type: Bug > Components: tests >Affects Versions: 1.10.9 >Reporter: Qian Yu >Priority: Major > > test_scheduler_job.py has a few flaky tests. Some are marked with > pytest.mark.xfail, but this one is not marked flaky. It sometimes fails in > Travis. For example: > > {code:python} > FAILURES > > _ > TestDagFileProcessor.test_dag_file_processor_process_task_instances_depends_on_past_0 > __ > a = ( testMethod=test_dag_file_processor_process_task_instances_depends_on_past_0>,) > @wraps(func) > def standalone_func(*a): > > return func(*(a + p.args), **p.kwargs) > /usr/local/lib/python3.6/site-packages/parameterized/parameterized.py:518: > _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ > _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ > self = testMethod=test_dag_file_processor_process_task_instances_depends_on_past_0> > state = None, start_date = None, end_date = None > @parameterized.expand([ > [State.NONE, None, None], > [State.UP_FOR_RETRY, timezone.utcnow() - > datetime.timedelta(minutes=30), > timezone.utcnow() - datetime.timedelta(minutes=15)], > [State.UP_FOR_RESCHEDULE, timezone.utcnow() - > datetime.timedelta(minutes=30), > timezone.utcnow() - datetime.timedelta(minutes=15)], > ]) > def test_dag_file_processor_process_task_instances_depends_on_past(self, > state, start_date, end_date): > """ > Test if _process_task_instances puts the right task instances into the > mock_list. > """ > dag = DAG( > dag_id='test_scheduler_process_execute_task_depends_on_past', > start_date=DEFAULT_DATE, > default_args={ > 'depends_on_past': True, > }, > ) > dag_task1 = DummyOperator( > task_id='dummy1', > dag=dag, > owner='airflow') > dag_task2 = DummyOp
[GitHub] [airflow] yuqian90 opened a new pull request #7470: [AIRFLOW-6834] Fix flaky test_scheduler_job by sorting TaskInstance
yuqian90 opened a new pull request #7470: [AIRFLOW-6834] Fix flaky test_scheduler_job by sorting TaskInstance URL: https://github.com/apache/airflow/pull/7470 Fix a flaky test that fails because `TaskInstance` are returned in non-deterministic order: ``` test_scheduler_job.py::TestDagFileProcessor::test_dag_file_processor_process_task_instances_depends_on_past ``` --- Issue link: WILL BE INSERTED BY [boring-cyborg](https://github.com/kaxil/boring-cyborg) Make sure to mark the boxes below before creating PR: [x] - [x] Description above provides context of the change - [x] Commit message/PR title starts with `[AIRFLOW-]`. AIRFLOW- = JIRA ID* - [x] Unit tests coverage for changes (not needed for documentation changes) - [x] Commits follow "[How to write a good git commit message](http://chris.beams.io/posts/git-commit/)" - [x] Relevant documentation is updated including usage instructions. - [x] I will engage committers as explained in [Contribution Workflow Example](https://github.com/apache/airflow/blob/master/CONTRIBUTING.rst#contribution-workflow-example). * For document-only changes commit message can start with `[AIRFLOW-]`. --- In case of fundamental code change, Airflow Improvement Proposal ([AIP](https://cwiki.apache.org/confluence/display/AIRFLOW/Airflow+Improvements+Proposals)) is needed. In case of a new dependency, check compliance with the [ASF 3rd Party License Policy](https://www.apache.org/legal/resolved.html#category-x). In case of backwards incompatible changes please leave a note in [UPDATING.md](https://github.com/apache/airflow/blob/master/UPDATING.md). Read the [Pull Request Guidelines](https://github.com/apache/airflow/blob/master/CONTRIBUTING.rst#pull-request-guidelines) for more information. This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[GitHub] [airflow] codecov-io edited a comment on issue #6870: [AIRFLOW-0578] Check return code
codecov-io edited a comment on issue #6870: [AIRFLOW-0578] Check return code URL: https://github.com/apache/airflow/pull/6870#issuecomment-569120114 # [Codecov](https://codecov.io/gh/apache/airflow/pull/6870?src=pr&el=h1) Report > Merging [#6870](https://codecov.io/gh/apache/airflow/pull/6870?src=pr&el=desc) into [master](https://codecov.io/gh/apache/airflow/commit/3730c24c41470cd331c5109539ee2fa0c9f4e74a?src=pr&el=desc) will **increase** coverage by `0.88%`. > The diff coverage is `100%`. [![Impacted file tree graph](https://codecov.io/gh/apache/airflow/pull/6870/graphs/tree.svg?width=650&token=WdLKlKHOAU&height=150&src=pr)](https://codecov.io/gh/apache/airflow/pull/6870?src=pr&el=tree) ```diff @@Coverage Diff@@ ## master #6870 +/- ## = + Coverage 85.52% 86.4% +0.88% = Files 758 882 +124 Lines 39932 41541+1609 = + Hits34150 35892+1742 + Misses 57825649 -133 ``` | [Impacted Files](https://codecov.io/gh/apache/airflow/pull/6870?src=pr&el=tree) | Coverage Δ | | |---|---|---| | [airflow/utils/state.py](https://codecov.io/gh/apache/airflow/pull/6870/diff?src=pr&el=tree#diff-YWlyZmxvdy91dGlscy9zdGF0ZS5weQ==) | `96.55% <100%> (+0.25%)` | :arrow_up: | | [airflow/jobs/local\_task\_job.py](https://codecov.io/gh/apache/airflow/pull/6870/diff?src=pr&el=tree#diff-YWlyZmxvdy9qb2JzL2xvY2FsX3Rhc2tfam9iLnB5) | `90.8% <100%> (+1.06%)` | :arrow_up: | | [airflow/jobs/base\_job.py](https://codecov.io/gh/apache/airflow/pull/6870/diff?src=pr&el=tree#diff-YWlyZmxvdy9qb2JzL2Jhc2Vfam9iLnB5) | `91.15% <100%> (-1.05%)` | :arrow_down: | | [airflow/operators/postgres\_operator.py](https://codecov.io/gh/apache/airflow/pull/6870/diff?src=pr&el=tree#diff-YWlyZmxvdy9vcGVyYXRvcnMvcG9zdGdyZXNfb3BlcmF0b3IucHk=) | `0% <0%> (-100%)` | :arrow_down: | | [airflow/operators/mysql\_to\_hive.py](https://codecov.io/gh/apache/airflow/pull/6870/diff?src=pr&el=tree#diff-YWlyZmxvdy9vcGVyYXRvcnMvbXlzcWxfdG9faGl2ZS5weQ==) | `0% <0%> (-100%)` | :arrow_down: | | [...rflow/providers/apache/cassandra/sensors/record.py](https://codecov.io/gh/apache/airflow/pull/6870/diff?src=pr&el=tree#diff-YWlyZmxvdy9wcm92aWRlcnMvYXBhY2hlL2Nhc3NhbmRyYS9zZW5zb3JzL3JlY29yZC5weQ==) | `0% <0%> (-100%)` | :arrow_down: | | [...irflow/providers/apache/cassandra/sensors/table.py](https://codecov.io/gh/apache/airflow/pull/6870/diff?src=pr&el=tree#diff-YWlyZmxvdy9wcm92aWRlcnMvYXBhY2hlL2Nhc3NhbmRyYS9zZW5zb3JzL3RhYmxlLnB5) | `0% <0%> (-100%)` | :arrow_down: | | [airflow/contrib/operators/snowflake\_operator.py](https://codecov.io/gh/apache/airflow/pull/6870/diff?src=pr&el=tree#diff-YWlyZmxvdy9jb250cmliL29wZXJhdG9ycy9zbm93Zmxha2Vfb3BlcmF0b3IucHk=) | `0% <0%> (-95.84%)` | :arrow_down: | | [airflow/operators/s3\_to\_hive\_operator.py](https://codecov.io/gh/apache/airflow/pull/6870/diff?src=pr&el=tree#diff-YWlyZmxvdy9vcGVyYXRvcnMvczNfdG9faGl2ZV9vcGVyYXRvci5weQ==) | `0% <0%> (-93.97%)` | :arrow_down: | | [airflow/contrib/hooks/grpc\_hook.py](https://codecov.io/gh/apache/airflow/pull/6870/diff?src=pr&el=tree#diff-YWlyZmxvdy9jb250cmliL2hvb2tzL2dycGNfaG9vay5weQ==) | `0% <0%> (-91.94%)` | :arrow_down: | | ... and [979 more](https://codecov.io/gh/apache/airflow/pull/6870/diff?src=pr&el=tree-more) | | -- [Continue to review full report at Codecov](https://codecov.io/gh/apache/airflow/pull/6870?src=pr&el=continue). > **Legend** - [Click here to learn more](https://docs.codecov.io/docs/codecov-delta) > `Δ = absolute (impact)`, `ø = not affected`, `? = missing data` > Powered by [Codecov](https://codecov.io/gh/apache/airflow/pull/6870?src=pr&el=footer). Last update [3730c24...6ea63bd](https://codecov.io/gh/apache/airflow/pull/6870?src=pr&el=lastupdated). Read the [comment docs](https://docs.codecov.io/docs/pull-request-comments). This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[GitHub] [airflow] zhongjiajie commented on issue #7463: [AIRFLOW-XXXX] clarify dag_id parameter
zhongjiajie commented on issue #7463: [AIRFLOW-] clarify dag_id parameter URL: https://github.com/apache/airflow/pull/7463#issuecomment-588573087 @MichaelChirico Not only your PR title, but also your git commit message. ![](https://i.loli.net/2020/02/20/NMiCU1zudGvqhs6.png) you could change and then force push to restart validator This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[GitHub] [airflow] mik-laj edited a comment on issue #5177: [AIRFLOW-4084] Fix bug downloading incomplete logs from ElasticSearch
mik-laj edited a comment on issue #5177: [AIRFLOW-4084] Fix bug downloading incomplete logs from ElasticSearch URL: https://github.com/apache/airflow/pull/5177#issuecomment-557788417 I wonder why this change had to make changes to the ``views.py`` file. Why were the Task Handler not updated? In my opinion, we should check the existence of the ``download_logs`` key in the implementation of the handler logic and then disable the pagination mechanism. Now the abstraction of code is running away and it is possible that we are breaking other handlers because they expected different behavior. I am working on documentation for this class and if we revert this change we will be able to do it as follows. In my opinion this is the original behavior of this code. ```python class TaskHandler(logging.Handler, ABC): """ Handler that allows you to write and read information about a specific task. """ @abstractmethod def read( self, task_instance: TaskInstance, try_number: Optional[int] = None, metadata: Optional[Dict] = None ) -> Tuple[List[str], List[Dict]]: """ Read logs of given task instance. It supports log pagination. To do this, the first call to this function contains an empty metadata object. As a result, list of logs and list of metadata should be returned. The resulting metadata should contain the key ``end_of_logs``, which determines whether pagination should be continued. It is possible to return more metadata objects, but only the first is used, so you should always return a list with one item. The remaining keys in the dictionary are sent back to the method without changes, which means that if you add an additional key with a token or page number, you can expect that the key will be available in the next request for logs. If the metadata in the call contains the ``download_logs'' key, then full logs should be returned without pagination. :param task_instance: task instance object :param try_number: task instance try_number to read logs from. If None it returns all logs separated by try_number :param metadata: log metadata, can be used for steaming log reading and auto-tailing. :return: a list of logs and list of metadata objects. """ ... @abstractmethod def set_context(self, task_instance: TaskInstance) -> None: """ Provide task_instance context to airflow task handler. Different implementations provide different behavior. Examples of behavior are: * in the case of handlers writing to a file, it may start writing to another file; * for remote services, it can start adding labels to logs. This allows us to later search for logs for a single task :param task_instance: task instance object """ ... ``` I would like to point out that in this PR there is a duplicate mechanism of handling the case when try_numbers is empty, i.e. the log for all try numbers is downloaded. Previously it was part of the file_task_handler handler, and now it has been copied a second time to another place despite the es_task_handler extended from this handler. https://github.com/apache/airflow/blob/master/airflow/utils/log/file_task_handler.py#L155-L164 https://github.com/apache/airflow/blob/master/airflow/utils/log/es_task_handler.py#L36 https://github.com/apache/airflow/blob/master/airflow/www/views.py#L595-L599 This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[GitHub] [airflow] digger commented on issue #6371: [AIRFLOW-5691] Rewrite Dataproc operators to use python library
digger commented on issue #6371: [AIRFLOW-5691] Rewrite Dataproc operators to use python library URL: https://github.com/apache/airflow/pull/6371#issuecomment-588548518 @dossett, I don't know if that was intentional. I just shared my feedback on changes made in AIRFLOW-3211. In the company I work for we use Airflow and we had to patch Airflow 1.10.7 and 1.10.9, reverting changes from AIRFLOW-3211, in order to make the dataproc functionality work. This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[airflow-site] branch asf-site updated: Update - Thu Feb 20 00:19:54 UTC 2020
This is an automated email from the ASF dual-hosted git repository. kamilbregula pushed a commit to branch asf-site in repository https://gitbox.apache.org/repos/asf/airflow-site.git The following commit(s) were added to refs/heads/asf-site by this push: new 3ffa600 Update - Thu Feb 20 00:19:54 UTC 2020 3ffa600 is described below commit 3ffa6004f0a4a9b1e22574263453a40aef2be76c Author: Kamil Bregula AuthorDate: Thu Feb 20 00:19:55 2020 + Update - Thu Feb 20 00:19:54 UTC 2020 --- 404.html | 4 +- blog/airflow-survey/index.html | 4 +- blog/announcing-new-website/index.html | 4 +- .../index.html | 4 +- .../index.html | 4 +- .../index.html | 4 +- blog/index.html| 4 +- .../index.html | 4 +- blog/tags/community/index.html | 4 +- blog/tags/development/index.html | 4 +- blog/tags/documentation/index.html | 4 +- blog/tags/survey/index.html| 4 +- blog/tags/users/index.html | 4 +- categories/index.html | 4 +- community/index.html | 4 +- index.html | 32 ++--- install/index.html | 4 +- integration-logos/apache/cassandra-3.png | Bin 0 -> 87960 bytes integration-logos/apache/druid-1.png | Bin 0 -> 28442 bytes integration-logos/apache/hadoop.png| Bin 0 -> 52103 bytes integration-logos/apache/hive.png | Bin 0 -> 141920 bytes integration-logos/apache/pig.png | Bin 0 -> 112286 bytes integration-logos/apache/pinot.png | Bin 0 -> 26792 bytes integration-logos/apache/spark.png | Bin 0 -> 66895 bytes integration-logos/apache/sqoop.png | Bin 0 -> 42874 bytes integration-logos/aws/aws-batch_light...@4x.png| Bin 0 -> 7274 bytes integration-logos/aws/aws-glue_light...@4x.png | Bin 0 -> 4392 bytes integration-logos/aws/aws-lambda_light...@4x.png | Bin 0 -> 4262 bytes .../aws/amazon-athena_light...@4x.png | Bin 0 -> 10215 bytes .../aws/amazon-cloudwatch_light...@4x.png | Bin 0 -> 7248 bytes .../aws/amazon-dynamodb_light...@4x.png| Bin 0 -> 7570 bytes integration-logos/aws/amazon-ec2_light...@4x.png | Bin 0 -> 2134 bytes integration-logos/aws/amazon-emr_light...@4x.png | Bin 0 -> 9456 bytes .../amazon-kinesis-data-firehose_light...@4x.png | Bin 0 -> 5259 bytes .../aws/amazon-redshift_light...@4x.png| Bin 0 -> 6391 bytes .../aws/amazon-sagemaker_light...@4x.png | Bin 0 -> 7702 bytes ...simple-notification-service-sns_light...@4x.png | Bin 0 -> 8103 bytes ...amazon-simple-queue-service-sqs_light...@4x.png | Bin 0 -> 8759 bytes ...mazon-simple-storage-service-s3_light...@4x.png | Bin 0 -> 7252 bytes integration-logos/azure-logo.svg | 24 -- integration-logos/azure/Azure Cosmos DB.svg| 11 + integration-logos/azure/Azure Files.svg| 8 integration-logos/azure/Blob Storage.svg | 9 integration-logos/azure/Container Instances.svg| 9 integration-logos/azure/Data Lake Storage.svg | 37 +++ integration-logos/gcp/AI-Platform.png | Bin 0 -> 6284 bytes integration-logos/gcp/BigQuery.png | Bin 0 -> 6210 bytes integration-logos/gcp/Cloud-AutoML.png | Bin 0 -> 6100 bytes integration-logos/gcp/Cloud-Bigtable.png | Bin 0 -> 8346 bytes integration-logos/gcp/Cloud-Build.png | Bin 0 -> 7075 bytes integration-logos/gcp/Cloud-Dataflow.png | Bin 0 -> 7332 bytes integration-logos/gcp/Cloud-Dataproc.png | Bin 0 -> 7257 bytes integration-logos/gcp/Cloud-Datastore.png | Bin 0 -> 4355 bytes integration-logos/gcp/Cloud-Functions.png | Bin 0 -> 5004 bytes integration-logos/gcp/Cloud-Memorystore.png| Bin 0 -> 4184 bytes integration-logos/gcp/Cloud-NLP.png| Bin 0 -> 3824 bytes integration-logos/gcp/Cloud-PubSub.png | Bin 0 -> 6756 bytes integration-logos/gcp/Cloud-SQL.png| Bin 0 -> 6704 bytes integration-logos/gcp/Cloud-Spanner.png| Bin 0 -> 5940 bytes integration-logos/gcp/Cloud-Speech-to-Text.png | Bin 0 -> 3830 bytes integration-logos/gcp/Cloud-Storage.png| Bin 0 -> 4189 bytes integration-logos/gcp/Cloud-Tasks.png | Bin 0 -> 17170 bytes integration-logos/gcp/Cloud-Text-to-Speech.png | Bin 0 -> 5030 bytes integration-logos/gcp/Cloud-Translation-API.png| Bin 0 -> 6645 bytes .../g
[GitHub] [airflow-site] mik-laj merged pull request #245: Add logos to integrations list
mik-laj merged pull request #245: Add logos to integrations list URL: https://github.com/apache/airflow-site/pull/245 This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[airflow-site] branch master updated: Add logos to integrations list (#245)
This is an automated email from the ASF dual-hosted git repository. kamilbregula pushed a commit to branch master in repository https://gitbox.apache.org/repos/asf/airflow-site.git The following commit(s) were added to refs/heads/master by this push: new 8495412 Add logos to integrations list (#245) 8495412 is described below commit 8495412957c20862448eae671064125d8f887a4e Author: Louis Guitton AuthorDate: Thu Feb 20 01:10:02 2020 +0100 Add logos to integrations list (#245) --- landing-pages/site/assets/scss/_list-boxes.scss| 2 + .../integration-logos/apache/cassandra-3.png | Bin 0 -> 87960 bytes .../static/integration-logos/apache/druid-1.png| Bin 0 -> 28442 bytes .../static/integration-logos/apache/hadoop.png | Bin 0 -> 52103 bytes .../site/static/integration-logos/apache/hive.png | Bin 0 -> 141920 bytes .../site/static/integration-logos/apache/pig.png | Bin 0 -> 112286 bytes .../site/static/integration-logos/apache/pinot.png | Bin 0 -> 26792 bytes .../site/static/integration-logos/apache/spark.png | Bin 0 -> 66895 bytes .../site/static/integration-logos/apache/sqoop.png | Bin 0 -> 42874 bytes .../aws/aws-batch_light...@4x.png | Bin 0 -> 7274 bytes .../integration-logos/aws/aws-glue_light...@4x.png | Bin 0 -> 4392 bytes .../aws/aws-lambda_light...@4x.png | Bin 0 -> 4262 bytes .../aws/amazon-athena_light...@4x.png | Bin 0 -> 10215 bytes .../aws/amazon-cloudwatch_light...@4x.png | Bin 0 -> 7248 bytes .../aws/amazon-dynamodb_light...@4x.png| Bin 0 -> 7570 bytes .../aws/amazon-ec2_light...@4x.png | Bin 0 -> 2134 bytes .../aws/amazon-emr_light...@4x.png | Bin 0 -> 9456 bytes .../amazon-kinesis-data-firehose_light...@4x.png | Bin 0 -> 5259 bytes .../aws/amazon-redshift_light...@4x.png| Bin 0 -> 6391 bytes .../aws/amazon-sagemaker_light...@4x.png | Bin 0 -> 7702 bytes ...simple-notification-service-sns_light...@4x.png | Bin 0 -> 8103 bytes ...amazon-simple-queue-service-sqs_light...@4x.png | Bin 0 -> 8759 bytes ...mazon-simple-storage-service-s3_light...@4x.png | Bin 0 -> 7252 bytes .../site/static/integration-logos/azure-logo.svg | 24 -- .../integration-logos/azure/Azure Cosmos DB.svg| 11 + .../static/integration-logos/azure/Azure Files.svg | 8 .../integration-logos/azure/Blob Storage.svg | 9 .../azure/Container Instances.svg | 9 .../integration-logos/azure/Data Lake Storage.svg | 37 +++ .../static/integration-logos/gcp/AI-Platform.png | Bin 0 -> 6284 bytes .../site/static/integration-logos/gcp/BigQuery.png | Bin 0 -> 6210 bytes .../static/integration-logos/gcp/Cloud-AutoML.png | Bin 0 -> 6100 bytes .../integration-logos/gcp/Cloud-Bigtable.png | Bin 0 -> 8346 bytes .../static/integration-logos/gcp/Cloud-Build.png | Bin 0 -> 7075 bytes .../integration-logos/gcp/Cloud-Dataflow.png | Bin 0 -> 7332 bytes .../integration-logos/gcp/Cloud-Dataproc.png | Bin 0 -> 7257 bytes .../integration-logos/gcp/Cloud-Datastore.png | Bin 0 -> 4355 bytes .../integration-logos/gcp/Cloud-Functions.png | Bin 0 -> 5004 bytes .../integration-logos/gcp/Cloud-Memorystore.png| Bin 0 -> 4184 bytes .../static/integration-logos/gcp/Cloud-NLP.png | Bin 0 -> 3824 bytes .../static/integration-logos/gcp/Cloud-PubSub.png | Bin 0 -> 6756 bytes .../static/integration-logos/gcp/Cloud-SQL.png | Bin 0 -> 6704 bytes .../static/integration-logos/gcp/Cloud-Spanner.png | Bin 0 -> 5940 bytes .../integration-logos/gcp/Cloud-Speech-to-Text.png | Bin 0 -> 3830 bytes .../static/integration-logos/gcp/Cloud-Storage.png | Bin 0 -> 4189 bytes .../static/integration-logos/gcp/Cloud-Tasks.png | Bin 0 -> 17170 bytes .../integration-logos/gcp/Cloud-Text-to-Speech.png | Bin 0 -> 5030 bytes .../gcp/Cloud-Translation-API.png | Bin 0 -> 6645 bytes .../gcp/Cloud-Video-Intelligence-API.png | Bin 0 -> 5810 bytes .../integration-logos/gcp/Cloud-Vision-API.png | Bin 0 -> 5822 bytes .../integration-logos/gcp/Compute-Engine.png | Bin 0 -> 4817 bytes .../gcp/Key-Management-Service.png | Bin 0 -> 7056 bytes .../integration-logos/gcp/Kubernetes-Engine.png| Bin 0 -> 6967 bytes landing-pages/site/static/integrations.json| 52 + 54 files changed, 128 insertions(+), 24 deletions(-) diff --git a/landing-pages/site/assets/scss/_list-boxes.scss b/landing-pages/site/assets/scss/_list-boxes.scss index 30709b3..c65afab 100644 --- a/landing-pages/site/assets/scss/_list-boxes.scss +++ b/landing-pages/site/assets/scss/_list-boxes.scss @@ -192,6 +192,8 @@ $card-margin: 20px; margin: auto 0; filter: grayscale(1); opacity: 0.6; + max-width: 100%; + max-height: 100%; } &--name { diff --git a/landing-pages/site/static/integration-logos/
[GitHub] [airflow-site] mik-laj commented on issue #245: Add logos to integrations list
mik-laj commented on issue #245: Add logos to integrations list URL: https://github.com/apache/airflow-site/pull/245#issuecomment-588540722 @kaxil This person is no longer working on Airflow, but I have a response from the design team. > cool, only those gray squares would reduce a bit (but that they would still be centered), because now they are a bit close to the text It's "cool", so we can merge it. This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[GitHub] [airflow-site] mik-laj commented on issue #246: Expand the guide for contributors
mik-laj commented on issue #246: Expand the guide for contributors URL: https://github.com/apache/airflow-site/pull/246#issuecomment-588539848 Thanks @kaxil This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[GitHub] [airflow] mik-laj commented on a change in pull request #7163: [AIRFLOW-6542] add spark-on-k8s operator/hook/sensor
mik-laj commented on a change in pull request #7163: [AIRFLOW-6542] add spark-on-k8s operator/hook/sensor URL: https://github.com/apache/airflow/pull/7163#discussion_r381607848 ## File path: tests/test_project_structure.py ## @@ -36,6 +36,9 @@ 'tests/providers/apache/pig/operators/test_pig.py', 'tests/providers/apache/spark/hooks/test_spark_jdbc_script.py', 'tests/providers/cncf/kubernetes/operators/test_kubernetes_pod.py', + 'tests/providers/cncf/kubernetes/operators/test_spark_kubernetes_operator.py', +'tests/providers/cncf/kubernetes/hooks/test_kubernetes_hook.py', +'tests/providers/cncf/kubernetes/sensors/test_spark_kubernetes_sensor.py', Review comment: Why is this not possible? I think that shouldn't be a problem. This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[GitHub] [airflow] dossett commented on issue #6371: [AIRFLOW-5691] Rewrite Dataproc operators to use python library
dossett commented on issue #6371: [AIRFLOW-5691] Rewrite Dataproc operators to use python library URL: https://github.com/apache/airflow/pull/6371#issuecomment-588523405 @digger I take your points about the original change. Do you know if reverting that functionality was an intentional part of this PR? This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[GitHub] [airflow-site] kaxil commented on issue #245: Add logos to integrations list
kaxil commented on issue #245: Add logos to integrations list URL: https://github.com/apache/airflow-site/pull/245#issuecomment-588512075 Awesome work @louisguitton 🎉 @kgabryje Any comments? This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[airflow-site] branch master updated: Expand the guide for contributors (#246)
This is an automated email from the ASF dual-hosted git repository. kaxilnaik pushed a commit to branch master in repository https://gitbox.apache.org/repos/asf/airflow-site.git The following commit(s) were added to refs/heads/master by this push: new ed0cf1a Expand the guide for contributors (#246) ed0cf1a is described below commit ed0cf1a0c0432e7f3f72c35080fc8953e042402d Author: Kamil Breguła AuthorDate: Wed Feb 19 23:41:14 2020 +0100 Expand the guide for contributors (#246) --- CONTRIBUTE.md | 559 ++ site.sh | 2 +- 2 files changed, 452 insertions(+), 109 deletions(-) diff --git a/CONTRIBUTE.md b/CONTRIBUTE.md index ea6f089..5a0d671 100644 --- a/CONTRIBUTE.md +++ b/CONTRIBUTE.md @@ -17,9 +17,10 @@ under the License. --> -## General directory structure +Contributor Guide += -```bash +``` . ├── dist ├── docs-archive @@ -54,49 +55,120 @@ └── sphinx_airflow_theme ``` -## Working with the project +# Working with the project + +Work with the site and documentation requires that your computer be properly prepared. Most tasks can +be done by the site.sh script + +### Prerequisite Tasks + +The following applications must be installed to use the project: + +* git +* docker + +It is also worth adding SSH keys for the `github.com` server to trusted ones. It is necessary to clone repositories. You can do this using following command: +```bash +ssh-keyscan -t rsa -H github.com >> ~/.ssh/known_hosts +``` + +**Debian instalation** + +To install git on Debian, run the following command: +```bash +sudo apt install git -y +``` + +To install docker, run the following command: +```bash +curl -fsSL https://get.docker.com -o get-docker.sh && sh get-docker.sh +sudo usermod -aG docker $USER +``` + +Git must have commit author information configured, run these commands +```bash +git config --global user.email '' +git config --global user.name '' +``` + +### Static checks + +The project uses many static checks using fantastic [pre-commit](https://pre-commit.com/). Every change is checked on CI and if it does not pass the tests it cannot be accepted. If you want to check locally then you should install Python3.6 or newer together with pip and run following command to install pre-commit: + +```bash +pip install -r requirements.txt +``` + +To turn on pre-commit checks for commit operations in git, enter: +```bash +pre-commit install +``` + +To run all checks on your staged files, enter: +```bash +pre-commit run +``` + +To run all checks on all files, enter: +```bash +pre-commit run --all-files +``` + +Pre-commit check results are also attached to your PR through integration with Travis CI. + +### Clone repository + +To clone repository from github.com to local disk, run following command + +```bash +git clone g...@github.com:apache/airflow-site.git +git submodule update --init --recursive +``` + +### Use `site.sh` script In order to run an environment for the project, make sure that you have Docker installed. Then, use the `site.sh` script to work with the website in a Docker container. `site.sh` provides the following commands. -build-site Prepare dist directory with landing pages and documentation -preview-siteStarts the web server with preview of the website -build-landing-pages Builds a landing pages -prepare-theme Prepares and copies files needed for the proper functioning of the sphinx theme. -shell Start shell -build-image Build a Docker image with a environment -install-node-deps Download all the Node dependencies -check-site-linksChecks if the links are correct in the website -lint-cssLint CSS files -lint-js Lint Javascript files -cleanup Delete the virtual environment in Docker -stopStop the environment -helpDisplay usage +build-sitePrepare dist directory with landing pages and documentation +preview-landing-pages Starts the web server with preview of the website +build-landing-pages Builds a landing pages +prepare-theme Prepares and copies files needed for the proper functioning of the sphinx theme. +shell Start shell +build-image Build a Docker image with a environment +install-node-deps Download all the Node dependencies +check-site-links Checks if the links are correct in the website +lint-css Lint CSS files +lint-js Lint Javascript files +cleanup Delete the virtual environment in Docker +stop Stop the environment +help Display usage ### How to add a new blogpost To add a new blogpost with pre-filled frontmatter, in `/landing-pages/site` run: - -hugo new blog/my-new-blogpost
[GitHub] [airflow-site] kaxil merged pull request #246: Expand the guide for contributors
kaxil merged pull request #246: Expand the guide for contributors URL: https://github.com/apache/airflow-site/pull/246 This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[GitHub] [airflow] roitvt commented on a change in pull request #7163: [AIRFLOW-6542] add spark-on-k8s operator/hook/sensor
roitvt commented on a change in pull request #7163: [AIRFLOW-6542] add spark-on-k8s operator/hook/sensor URL: https://github.com/apache/airflow/pull/7163#discussion_r381585525 ## File path: airflow/providers/cncf/kubernetes/operators/spark_kubernetes_operator.py ## @@ -0,0 +1,83 @@ +# +# Licensed to the Apache Software Foundation (ASF) under one +# or more contributor license agreements. See the NOTICE file +# distributed with this work for additional information +# regarding copyright ownership. The ASF licenses this file +# to you under the Apache License, Version 2.0 (the +# "License"); you may not use this file except in compliance +# with the License. You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, +# software distributed under the License is distributed on an +# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY +# KIND, either express or implied. See the License for the +# specific language governing permissions and limitations +# under the License. +from typing import Optional + +import yaml +from kubernetes import client + +from airflow.exceptions import AirflowException +from airflow.models import BaseOperator +from airflow.providers.cncf.kubernetes.hooks.kubernetes_hook import Kuberneteshook +from airflow.utils.decorators import apply_defaults + + +class SparkKubernetesOperator(BaseOperator): +""" +Creates sparkApplication object in kubernetes cluster: + .. seealso:: +For more detail about Spark Application Object have a look at the reference: + https://github.com/GoogleCloudPlatform/spark-on-k8s-operator/blob/master/docs/api-docs.md#sparkapplication + +:param sparkapplication_file: filepath to kubernetes custom_resource_definition of sparkApplication +:type sparkapplication_file: str +:param namespace: kubernetes namespace to put sparkApplication +:type namespace: str +:param conn_id: the connection to Kubernetes cluster +:type conn_id: str +""" + +template_fields = ['sparkapplication_file', 'namespace'] +template_ext = ('yaml', 'yml', 'json') +ui_color = '#f4a460' + +@apply_defaults +def __init__(self, + sparkapplication_file: str, + namespace: Optional[str] = None, + conn_id: str = 'kubernetes_default', + *args, **kwargs) -> None: +super().__init__(*args, **kwargs) +self.sparkapplication_file = sparkapplication_file +self.namespace = namespace +self.conn_id = conn_id + +def execute(self, context): +self.log.info("Creating sparkApplication") +hook = Kuberneteshook(conn_id=self.conn_id) +api_client = hook.get_conn() +api = client.CustomObjectsApi(api_client) Review comment: but then I'll need to write code in the hook for every Kubernetes API instead of just return general API connection and the operator decide which API kind he opens. what do you think is right? This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[GitHub] [airflow] mik-laj commented on issue #7458: [AIRFLOW-6838][WIP] Introduce real subcommands for Breeze
mik-laj commented on issue #7458: [AIRFLOW-6838][WIP] Introduce real subcommands for Breeze URL: https://github.com/apache/airflow/pull/7458#issuecomment-588507636 > they're still there. pre-commit got stuck when asking to rebuild the image. I pushed the changes now. > Not necessary. that's just splitting help - nothing else. One parser should be enough. Again - happy to add it. I invite you to contribute. This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[GitHub] [airflow] roitvt commented on a change in pull request #7163: [AIRFLOW-6542] add spark-on-k8s operator/hook/sensor
roitvt commented on a change in pull request #7163: [AIRFLOW-6542] add spark-on-k8s operator/hook/sensor URL: https://github.com/apache/airflow/pull/7163#discussion_r381583482 ## File path: tests/test_project_structure.py ## @@ -36,6 +36,9 @@ 'tests/providers/apache/pig/operators/test_pig.py', 'tests/providers/apache/spark/hooks/test_spark_jdbc_script.py', 'tests/providers/cncf/kubernetes/operators/test_kubernetes_pod.py', + 'tests/providers/cncf/kubernetes/operators/test_spark_kubernetes_operator.py', +'tests/providers/cncf/kubernetes/hooks/test_kubernetes_hook.py', +'tests/providers/cncf/kubernetes/sensors/test_spark_kubernetes_sensor.py', Review comment: I know :( unfortunately it still not possible to mock Kubernetes API in python for unit testing but I'll try to write tests to what I can. This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[GitHub] [airflow] mik-laj commented on a change in pull request #7450: [AIRFLOW-6829][WIP] Introduce BaseOperatorMetaClass - auto-apply apply_default
mik-laj commented on a change in pull request #7450: [AIRFLOW-6829][WIP] Introduce BaseOperatorMetaClass - auto-apply apply_default URL: https://github.com/apache/airflow/pull/7450#discussion_r381580741 ## File path: airflow/operators/check_operator.py ## @@ -123,11 +120,10 @@ class ValueCheckOperator(BaseOperator): __mapper_args__ = { 'polymorphic_identity': 'ValueCheckOperator' } -template_fields = ('sql', 'pass_value',) # type: Iterable[str] +template_fields = ('sql', 'pass_value',) # type: Tuple[str, ...] Review comment: mypy is going crazy in many places now. It was one of the attempts to fix it. This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[GitHub] [airflow] roitvt commented on a change in pull request #7163: [AIRFLOW-6542] add spark-on-k8s operator/hook/sensor
roitvt commented on a change in pull request #7163: [AIRFLOW-6542] add spark-on-k8s operator/hook/sensor URL: https://github.com/apache/airflow/pull/7163#discussion_r381580906 ## File path: airflow/providers/cncf/kubernetes/hooks/kubernetes_hook.py ## @@ -0,0 +1,77 @@ +# Licensed to the Apache Software Foundation (ASF) under one +# or more contributor license agreements. See the NOTICE file +# distributed with this work for additional information +# regarding copyright ownership. The ASF licenses this file +# to you under the Apache License, Version 2.0 (the +# "License"); you may not use this file except in compliance +# with the License. You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, +# software distributed under the License is distributed on an +# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY +# KIND, either express or implied. See the License for the +# specific language governing permissions and limitations +# under the License. +import tempfile + +from kubernetes import client, config + +from airflow.hooks.base_hook import BaseHook + + +class Kuberneteshook(BaseHook): Review comment: so would you like me to open new PR for the Kubernetes connection? I think that KubePodOperator can benefit from using this hook as it can make it work on remote k8s clusters. This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[GitHub] [airflow] mik-laj commented on a change in pull request #7450: [AIRFLOW-6829][WIP] Introduce BaseOperatorMetaClass - auto-apply apply_default
mik-laj commented on a change in pull request #7450: [AIRFLOW-6829][WIP] Introduce BaseOperatorMetaClass - auto-apply apply_default URL: https://github.com/apache/airflow/pull/7450#discussion_r381580741 ## File path: airflow/operators/check_operator.py ## @@ -123,11 +120,10 @@ class ValueCheckOperator(BaseOperator): __mapper_args__ = { 'polymorphic_identity': 'ValueCheckOperator' } -template_fields = ('sql', 'pass_value',) # type: Iterable[str] +template_fields = ('sql', 'pass_value',) # type: Tuple[str, ...] Review comment: mypy is going crazy in many places now. It was one of the attempts to do it. This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[jira] [Commented] (AIRFLOW-6847) Integrate Apache Hive tests with Breeze
[ https://issues.apache.org/jira/browse/AIRFLOW-6847?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17040477#comment-17040477 ] Kamil Bregula commented on AIRFLOW-6847: I updated the title and the description to better describe the necessary scope of work. > Integrate Apache Hive tests with Breeze > --- > > Key: AIRFLOW-6847 > URL: https://issues.apache.org/jira/browse/AIRFLOW-6847 > Project: Apache Airflow > Issue Type: Improvement > Components: tests >Affects Versions: 1.10.9 >Reporter: Cooper Gillan >Priority: Minor > > Currently, tests for Apache Hive are not run on CI. This is very > sad.:crying_cat_face: > However, the tests exist in the repository and can be run using > AIRFLOW_RUNALL_TESTS environment variable. > I think that to solve this problem we need to follow these steps. > # Add Hive integration in Breeze > # Replace AIRFLOW_RUNALL_TESTS env variable with pytest marker > # Update .travis.yml > > This problem was found while working on splitting out Hive tests captured in > AIRFLOW-6721 and initially worked on in > [https://github.com/apache/airflow/pull/7468], > More info on airflow's use of Breeze integration and pytest markers can be > found here: > [https://github.com/apache/airflow/blob/master/TESTING.rst#airflow-integration-tests] > This ticket rose out of [this comment in PR #7468's > review|https://github.com/apache/airflow/pull/7468#issuecomment-588458850]. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Updated] (AIRFLOW-6847) Integrate Apache Hive tests with Breeze
[ https://issues.apache.org/jira/browse/AIRFLOW-6847?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Kamil Bregula updated AIRFLOW-6847: --- Description: Currently, tests for Apache Hive are not run on CI. This is very sad.:crying_cat_face: However, the tests exist in the repository and can be run using AIRFLOW_RUNALL_TESTS environment variable. I think that to solve this problem we need to follow these steps. # Add Hive integration in Breeze # Replace AIRFLOW_RUNALL_TESTS env variable with pytest marker # Update .travis.yml This problem was found while working on splitting out Hive tests captured in AIRFLOW-6721 and initially worked on in [https://github.com/apache/airflow/pull/7468], More info on airflow's use of Breeze integration and pytest markers can be found here: [https://github.com/apache/airflow/blob/master/TESTING.rst#airflow-integration-tests] This ticket rose out of [this comment in PR #7468's review|https://github.com/apache/airflow/pull/7468#issuecomment-588458850]. was: Currently, tests for Apache Hive are not run on CI. This is very sad.:crying_cat_face: However, the tests exist in the repository and can be run using AIRFLOW_RUNALL_TESTS environment variable. I think that to solve this problem we need to follow these steps. # Add Hive integration in Breeze # Replace AIRFLOW_RUNALL_TESTS env variable use of with pytest marker # Update .travis.yml This problem was found while working on splitting out Hive tests captured in AIRFLOW-6721 and initially worked on in [https://github.com/apache/airflow/pull/7468], More info on airflow's use of Breeze integration and pytest markers can be found here: [https://github.com/apache/airflow/blob/master/TESTING.rst#airflow-integration-tests] This ticket rose out of [this comment in PR #7468's review|https://github.com/apache/airflow/pull/7468#issuecomment-588458850]. > Integrate Apache Hive tests with Breeze > --- > > Key: AIRFLOW-6847 > URL: https://issues.apache.org/jira/browse/AIRFLOW-6847 > Project: Apache Airflow > Issue Type: Improvement > Components: tests >Affects Versions: 1.10.9 >Reporter: Cooper Gillan >Priority: Minor > > Currently, tests for Apache Hive are not run on CI. This is very > sad.:crying_cat_face: > However, the tests exist in the repository and can be run using > AIRFLOW_RUNALL_TESTS environment variable. > I think that to solve this problem we need to follow these steps. > # Add Hive integration in Breeze > # Replace AIRFLOW_RUNALL_TESTS env variable with pytest marker > # Update .travis.yml > > This problem was found while working on splitting out Hive tests captured in > AIRFLOW-6721 and initially worked on in > [https://github.com/apache/airflow/pull/7468], > More info on airflow's use of Breeze integration and pytest markers can be > found here: > [https://github.com/apache/airflow/blob/master/TESTING.rst#airflow-integration-tests] > This ticket rose out of [this comment in PR #7468's > review|https://github.com/apache/airflow/pull/7468#issuecomment-588458850]. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[GitHub] [airflow] roitvt commented on a change in pull request #7163: [AIRFLOW-6542] add spark-on-k8s operator/hook/sensor
roitvt commented on a change in pull request #7163: [AIRFLOW-6542] add spark-on-k8s operator/hook/sensor URL: https://github.com/apache/airflow/pull/7163#discussion_r381579048 ## File path: airflow/providers/cncf/kubernetes/hooks/kubernetes_hook.py ## @@ -0,0 +1,77 @@ +# Licensed to the Apache Software Foundation (ASF) under one +# or more contributor license agreements. See the NOTICE file +# distributed with this work for additional information +# regarding copyright ownership. The ASF licenses this file +# to you under the Apache License, Version 2.0 (the +# "License"); you may not use this file except in compliance +# with the License. You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, +# software distributed under the License is distributed on an +# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY +# KIND, either express or implied. See the License for the +# specific language governing permissions and limitations +# under the License. +import tempfile + +from kubernetes import client, config + +from airflow.hooks.base_hook import BaseHook + + +class Kuberneteshook(BaseHook): +""" +Creates Kubernetes API connection. + +:param conn_id: the connection to Kubernetes cluster +""" + +def __init__( +self, +conn_id="kubernetes_default" +): +self.connection = self.get_connection(conn_id) +self.extras = self.connection.extra_dejson + +def get_conn(self): +""" +Returns kubernetes api session for use with requests +""" + +if self._get_field(("in_cluster")): +self.log.debug("loading kube_config from: in_cluster configuration") +config.load_incluster_config() +elif self._get_field("kube_config") is None or self._get_field("kube_config") == '': +self.log.debug("loading kube_config from: default file") +else: +with tempfile.NamedTemporaryFile() as temp_config: +self.log.debug("loading kube_config from: connection kube_config") +temp_config.write(self._get_field("kube_config").encode()) +config.load_kube_config(temp_config.name) +temp_config.flush() Review comment: I don't like this solution either, but I dug into Kubernetes python API and didn't found a better solution. so I write the Kube config to temp file let Kubernetes python API to load it and then flush. This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[jira] [Updated] (AIRFLOW-6847) Integrate Apache Hive tests with Breeze
[ https://issues.apache.org/jira/browse/AIRFLOW-6847?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Kamil Bregula updated AIRFLOW-6847: --- Description: Currently, tests for Apache Hive are not run on CI. This is very sad.:crying_cat_face: However, the tests exist in the repository and can be run using AIRFLOW_RUNALL_TESTS environment variable. I think that to solve this problem we need to follow these steps. # Add Hive integration in Breeze # Replace AIRFLOW_RUNALL_TESTS env variable use of with pytest marker # Update .travis.yml This problem was found while working on splitting out Hive tests captured in AIRFLOW-6721 and initially worked on in [https://github.com/apache/airflow/pull/7468], More info on airflow's use of Breeze integration and pytest markers can be found here: [https://github.com/apache/airflow/blob/master/TESTING.rst#airflow-integration-tests] This ticket rose out of [this comment in PR #7468's review|https://github.com/apache/airflow/pull/7468#issuecomment-588458850]. was: Building off work completed to split out Hive tests captured in AIRFLOW-6721 and initially worked on in https://github.com/apache/airflow/pull/7468, remove use of the {{AIRFLOW_RUNALL_TESTS}} environment variable for skipping certain tests in favor of using pytest markers. More info on airflow's use of pytest markers can be found here: https://github.com/apache/airflow/blob/master/TESTING.rst#airflow-integration-tests This ticket rose out of [this comment in PR #7468's review|https://github.com/apache/airflow/pull/7468#issuecomment-588458850]. Summary: Integrate Apache Hive tests with Breeze (was: Replace use of AIRFLOW_RUNALL_TESTS with pytest markers) > Integrate Apache Hive tests with Breeze > --- > > Key: AIRFLOW-6847 > URL: https://issues.apache.org/jira/browse/AIRFLOW-6847 > Project: Apache Airflow > Issue Type: Improvement > Components: tests >Affects Versions: 1.10.9 >Reporter: Cooper Gillan >Priority: Minor > > Currently, tests for Apache Hive are not run on CI. This is very > sad.:crying_cat_face: > However, the tests exist in the repository and can be run using > AIRFLOW_RUNALL_TESTS environment variable. > I think that to solve this problem we need to follow these steps. > # Add Hive integration in Breeze > # Replace AIRFLOW_RUNALL_TESTS env variable use of with pytest marker > # Update .travis.yml > > This problem was found while working on splitting out Hive tests captured in > AIRFLOW-6721 and initially worked on in > [https://github.com/apache/airflow/pull/7468], > More info on airflow's use of Breeze integration and pytest markers can be > found here: > [https://github.com/apache/airflow/blob/master/TESTING.rst#airflow-integration-tests] > This ticket rose out of [this comment in PR #7468's > review|https://github.com/apache/airflow/pull/7468#issuecomment-588458850]. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Commented] (AIRFLOW-6847) Replace use of AIRFLOW_RUNALL_TESTS with pytest markers
[ https://issues.apache.org/jira/browse/AIRFLOW-6847?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17040472#comment-17040472 ] Kamil Bregula commented on AIRFLOW-6847: It is likely that these integrations should have a separate Apache Hive container running. This will allow these tests to be run and tested more easily. Now these tests are not run by anyone. :crying_cat_face: This serious problem is for these integrations. > Replace use of AIRFLOW_RUNALL_TESTS with pytest markers > --- > > Key: AIRFLOW-6847 > URL: https://issues.apache.org/jira/browse/AIRFLOW-6847 > Project: Apache Airflow > Issue Type: Improvement > Components: tests >Affects Versions: 1.10.9 >Reporter: Cooper Gillan >Priority: Minor > > Building off work completed to split out Hive tests captured in AIRFLOW-6721 > and initially worked on in https://github.com/apache/airflow/pull/7468, > remove use of the {{AIRFLOW_RUNALL_TESTS}} environment variable for skipping > certain tests in favor of using pytest markers. > More info on airflow's use of pytest markers can be found here: > https://github.com/apache/airflow/blob/master/TESTING.rst#airflow-integration-tests > This ticket rose out of [this comment in PR #7468's > review|https://github.com/apache/airflow/pull/7468#issuecomment-588458850]. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Commented] (AIRFLOW-6847) Replace use of AIRFLOW_RUNALL_TESTS with pytest markers
[ https://issues.apache.org/jira/browse/AIRFLOW-6847?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17040468#comment-17040468 ] Cooper Gillan commented on AIRFLOW-6847: [~kamil.bregula] - which integration would these Hive tests be using? Or are we thinking for a start we can/should just use {{@pytest.mark.skipif}}? > Replace use of AIRFLOW_RUNALL_TESTS with pytest markers > --- > > Key: AIRFLOW-6847 > URL: https://issues.apache.org/jira/browse/AIRFLOW-6847 > Project: Apache Airflow > Issue Type: Improvement > Components: tests >Affects Versions: 1.10.9 >Reporter: Cooper Gillan >Priority: Minor > > Building off work completed to split out Hive tests captured in AIRFLOW-6721 > and initially worked on in https://github.com/apache/airflow/pull/7468, > remove use of the {{AIRFLOW_RUNALL_TESTS}} environment variable for skipping > certain tests in favor of using pytest markers. > More info on airflow's use of pytest markers can be found here: > https://github.com/apache/airflow/blob/master/TESTING.rst#airflow-integration-tests > This ticket rose out of [this comment in PR #7468's > review|https://github.com/apache/airflow/pull/7468#issuecomment-588458850]. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[GitHub] [airflow] kujon commented on a change in pull request #7163: [AIRFLOW-6542] add spark-on-k8s operator/hook/sensor
kujon commented on a change in pull request #7163: [AIRFLOW-6542] add spark-on-k8s operator/hook/sensor URL: https://github.com/apache/airflow/pull/7163#discussion_r381572792 ## File path: airflow/providers/cncf/kubernetes/operators/spark_kubernetes_operator.py ## @@ -0,0 +1,83 @@ +# +# Licensed to the Apache Software Foundation (ASF) under one +# or more contributor license agreements. See the NOTICE file +# distributed with this work for additional information +# regarding copyright ownership. The ASF licenses this file +# to you under the Apache License, Version 2.0 (the +# "License"); you may not use this file except in compliance +# with the License. You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, +# software distributed under the License is distributed on an +# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY +# KIND, either express or implied. See the License for the +# specific language governing permissions and limitations +# under the License. +from typing import Optional + +import yaml +from kubernetes import client + +from airflow.exceptions import AirflowException +from airflow.models import BaseOperator +from airflow.providers.cncf.kubernetes.hooks.kubernetes_hook import Kuberneteshook +from airflow.utils.decorators import apply_defaults + + +class SparkKubernetesOperator(BaseOperator): +""" +Creates sparkApplication object in kubernetes cluster: + .. seealso:: +For more detail about Spark Application Object have a look at the reference: + https://github.com/GoogleCloudPlatform/spark-on-k8s-operator/blob/master/docs/api-docs.md#sparkapplication + +:param sparkapplication_file: filepath to kubernetes custom_resource_definition of sparkApplication +:type sparkapplication_file: str +:param namespace: kubernetes namespace to put sparkApplication +:type namespace: str +:param conn_id: the connection to Kubernetes cluster +:type conn_id: str +""" + +template_fields = ['sparkapplication_file', 'namespace'] +template_ext = ('yaml', 'yml', 'json') +ui_color = '#f4a460' + +@apply_defaults +def __init__(self, + sparkapplication_file: str, Review comment: apologies, I missed that! in such case, the only thing then is the inconsistency with the `KubernetesPodOperator`, all of the functionality is there. This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[GitHub] [airflow] potiuk edited a comment on issue #7458: [AIRFLOW-6838][WIP] Introduce real subcommands for Breeze
potiuk edited a comment on issue #7458: [AIRFLOW-6838][WIP] Introduce real subcommands for Breeze URL: https://github.com/apache/airflow/pull/7458#issuecomment-588491983 > This is a big change, because it will also require that we do not have one parser, but several smaller parsers. Not necessary. that's just splitting help - nothing else. One parser should be enough. Again - happy to add it. This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[GitHub] [airflow] potiuk commented on issue #7458: [AIRFLOW-6838][WIP] Introduce real subcommands for Breeze
potiuk commented on issue #7458: [AIRFLOW-6838][WIP] Introduce real subcommands for Breeze URL: https://github.com/apache/airflow/pull/7458#issuecomment-588491983 > This is a big change, because it will also require that we do not have one parser, but several smaller parsers. Not necessary. that's just splitting help - nothing else. One parser should be enough This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[GitHub] [airflow] potiuk commented on issue #7458: [AIRFLOW-6838][WIP] Introduce real subcommands for Breeze
potiuk commented on issue #7458: [AIRFLOW-6838][WIP] Introduce real subcommands for Breeze URL: https://github.com/apache/airflow/pull/7458#issuecomment-588491335 > > I think we should get rid of the short one-letter options for sub-commands. > > Removed they're still there. This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[GitHub] [airflow] kaxil commented on a change in pull request #7163: [AIRFLOW-6542] add spark-on-k8s operator/hook/sensor
kaxil commented on a change in pull request #7163: [AIRFLOW-6542] add spark-on-k8s operator/hook/sensor URL: https://github.com/apache/airflow/pull/7163#discussion_r381565465 ## File path: airflow/providers/cncf/kubernetes/hooks/kubernetes_hook.py ## @@ -0,0 +1,77 @@ +# Licensed to the Apache Software Foundation (ASF) under one +# or more contributor license agreements. See the NOTICE file +# distributed with this work for additional information +# regarding copyright ownership. The ASF licenses this file +# to you under the Apache License, Version 2.0 (the +# "License"); you may not use this file except in compliance +# with the License. You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, +# software distributed under the License is distributed on an +# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY +# KIND, either express or implied. See the License for the +# specific language governing permissions and limitations +# under the License. +import tempfile + +from kubernetes import client, config + +from airflow.hooks.base_hook import BaseHook + + +class Kuberneteshook(BaseHook): Review comment: Should be a separate PR though This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[GitHub] [airflow] kaxil commented on a change in pull request #7163: [AIRFLOW-6542] add spark-on-k8s operator/hook/sensor
kaxil commented on a change in pull request #7163: [AIRFLOW-6542] add spark-on-k8s operator/hook/sensor URL: https://github.com/apache/airflow/pull/7163#discussion_r381565037 ## File path: airflow/providers/cncf/kubernetes/hooks/kubernetes_hook.py ## @@ -0,0 +1,77 @@ +# Licensed to the Apache Software Foundation (ASF) under one +# or more contributor license agreements. See the NOTICE file +# distributed with this work for additional information +# regarding copyright ownership. The ASF licenses this file +# to you under the Apache License, Version 2.0 (the +# "License"); you may not use this file except in compliance +# with the License. You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, +# software distributed under the License is distributed on an +# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY +# KIND, either express or implied. See the License for the +# specific language governing permissions and limitations +# under the License. +import tempfile + +from kubernetes import client, config + +from airflow.hooks.base_hook import BaseHook + + +class Kuberneteshook(BaseHook): Review comment: I feel like KubePodOperator should use this hook. This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[GitHub] [airflow] ashb commented on a change in pull request #7163: [AIRFLOW-6542] add spark-on-k8s operator/hook/sensor
ashb commented on a change in pull request #7163: [AIRFLOW-6542] add spark-on-k8s operator/hook/sensor URL: https://github.com/apache/airflow/pull/7163#discussion_r381563497 ## File path: airflow/providers/cncf/kubernetes/hooks/kubernetes_hook.py ## @@ -0,0 +1,77 @@ +# Licensed to the Apache Software Foundation (ASF) under one +# or more contributor license agreements. See the NOTICE file +# distributed with this work for additional information +# regarding copyright ownership. The ASF licenses this file +# to you under the Apache License, Version 2.0 (the +# "License"); you may not use this file except in compliance +# with the License. You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, +# software distributed under the License is distributed on an +# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY +# KIND, either express or implied. See the License for the +# specific language governing permissions and limitations +# under the License. +import tempfile + +from kubernetes import client, config + +from airflow.hooks.base_hook import BaseHook + + +class Kuberneteshook(BaseHook): Review comment: I wonder if we could share some of the connection code we already have for both the KubernetesPodOperator and the Kubernetes Executor. Or should the KubePodOperator be slowly reworked (in a separate PR) to use this hook? This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[GitHub] [airflow] ashb commented on a change in pull request #7163: [AIRFLOW-6542] add spark-on-k8s operator/hook/sensor
ashb commented on a change in pull request #7163: [AIRFLOW-6542] add spark-on-k8s operator/hook/sensor URL: https://github.com/apache/airflow/pull/7163#discussion_r381561596 ## File path: tests/test_project_structure.py ## @@ -36,6 +36,9 @@ 'tests/providers/apache/pig/operators/test_pig.py', 'tests/providers/apache/spark/hooks/test_spark_jdbc_script.py', 'tests/providers/cncf/kubernetes/operators/test_kubernetes_pod.py', + 'tests/providers/cncf/kubernetes/operators/test_spark_kubernetes_operator.py', +'tests/providers/cncf/kubernetes/hooks/test_kubernetes_hook.py', +'tests/providers/cncf/kubernetes/sensors/test_spark_kubernetes_sensor.py', Review comment: Naughty Naughty. This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[GitHub] [airflow] ashb commented on a change in pull request #7163: [AIRFLOW-6542] add spark-on-k8s operator/hook/sensor
ashb commented on a change in pull request #7163: [AIRFLOW-6542] add spark-on-k8s operator/hook/sensor URL: https://github.com/apache/airflow/pull/7163#discussion_r381560358 ## File path: airflow/providers/cncf/kubernetes/operators/spark_kubernetes_operator.py ## @@ -0,0 +1,83 @@ +# +# Licensed to the Apache Software Foundation (ASF) under one +# or more contributor license agreements. See the NOTICE file +# distributed with this work for additional information +# regarding copyright ownership. The ASF licenses this file +# to you under the Apache License, Version 2.0 (the +# "License"); you may not use this file except in compliance +# with the License. You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, +# software distributed under the License is distributed on an +# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY +# KIND, either express or implied. See the License for the +# specific language governing permissions and limitations +# under the License. +from typing import Optional + +import yaml +from kubernetes import client + +from airflow.exceptions import AirflowException +from airflow.models import BaseOperator +from airflow.providers.cncf.kubernetes.hooks.kubernetes_hook import Kuberneteshook +from airflow.utils.decorators import apply_defaults + + +class SparkKubernetesOperator(BaseOperator): +""" +Creates sparkApplication object in kubernetes cluster: + .. seealso:: +For more detail about Spark Application Object have a look at the reference: + https://github.com/GoogleCloudPlatform/spark-on-k8s-operator/blob/master/docs/api-docs.md#sparkapplication + +:param sparkapplication_file: filepath to kubernetes custom_resource_definition of sparkApplication +:type sparkapplication_file: str +:param namespace: kubernetes namespace to put sparkApplication +:type namespace: str +:param conn_id: the connection to Kubernetes cluster +:type conn_id: str +""" + +template_fields = ['sparkapplication_file', 'namespace'] +template_ext = ('yaml', 'yml', 'json') +ui_color = '#f4a460' + +@apply_defaults +def __init__(self, + sparkapplication_file: str, + namespace: Optional[str] = None, + conn_id: str = 'kubernetes_default', + *args, **kwargs) -> None: +super().__init__(*args, **kwargs) +self.sparkapplication_file = sparkapplication_file +self.namespace = namespace +self.conn_id = conn_id + +def execute(self, context): +self.log.info("Creating sparkApplication") +hook = Kuberneteshook(conn_id=self.conn_id) +api_client = hook.get_conn() +api = client.CustomObjectsApi(api_client) +try: +sparkapplication_dict = yaml.safe_load(self.sparkapplication_file) Review comment: Since you have specified the template_ext, if this file exists the contents of the `sparkapplication_file` attribute will be _set to the content of the file_, not the filename. This needs tests! Edit: which won't break the operator (I'm confusing yaml with json which has load and loads methods). Still. This definitely needs tests added to the PR. This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[GitHub] [airflow] ashb commented on a change in pull request #7163: [AIRFLOW-6542] add spark-on-k8s operator/hook/sensor
ashb commented on a change in pull request #7163: [AIRFLOW-6542] add spark-on-k8s operator/hook/sensor URL: https://github.com/apache/airflow/pull/7163#discussion_r381560358 ## File path: airflow/providers/cncf/kubernetes/operators/spark_kubernetes_operator.py ## @@ -0,0 +1,83 @@ +# +# Licensed to the Apache Software Foundation (ASF) under one +# or more contributor license agreements. See the NOTICE file +# distributed with this work for additional information +# regarding copyright ownership. The ASF licenses this file +# to you under the Apache License, Version 2.0 (the +# "License"); you may not use this file except in compliance +# with the License. You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, +# software distributed under the License is distributed on an +# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY +# KIND, either express or implied. See the License for the +# specific language governing permissions and limitations +# under the License. +from typing import Optional + +import yaml +from kubernetes import client + +from airflow.exceptions import AirflowException +from airflow.models import BaseOperator +from airflow.providers.cncf.kubernetes.hooks.kubernetes_hook import Kuberneteshook +from airflow.utils.decorators import apply_defaults + + +class SparkKubernetesOperator(BaseOperator): +""" +Creates sparkApplication object in kubernetes cluster: + .. seealso:: +For more detail about Spark Application Object have a look at the reference: + https://github.com/GoogleCloudPlatform/spark-on-k8s-operator/blob/master/docs/api-docs.md#sparkapplication + +:param sparkapplication_file: filepath to kubernetes custom_resource_definition of sparkApplication +:type sparkapplication_file: str +:param namespace: kubernetes namespace to put sparkApplication +:type namespace: str +:param conn_id: the connection to Kubernetes cluster +:type conn_id: str +""" + +template_fields = ['sparkapplication_file', 'namespace'] +template_ext = ('yaml', 'yml', 'json') +ui_color = '#f4a460' + +@apply_defaults +def __init__(self, + sparkapplication_file: str, + namespace: Optional[str] = None, + conn_id: str = 'kubernetes_default', + *args, **kwargs) -> None: +super().__init__(*args, **kwargs) +self.sparkapplication_file = sparkapplication_file +self.namespace = namespace +self.conn_id = conn_id + +def execute(self, context): +self.log.info("Creating sparkApplication") +hook = Kuberneteshook(conn_id=self.conn_id) +api_client = hook.get_conn() +api = client.CustomObjectsApi(api_client) +try: +sparkapplication_dict = yaml.safe_load(self.sparkapplication_file) Review comment: Since you have specified the template_ext, if this file exists the contents of the `sparkapplication_file` attribute will be _set to the content of the file_, not the filename. This needs tests! This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[GitHub] [airflow] ashb commented on a change in pull request #7163: [AIRFLOW-6542] add spark-on-k8s operator/hook/sensor
ashb commented on a change in pull request #7163: [AIRFLOW-6542] add spark-on-k8s operator/hook/sensor URL: https://github.com/apache/airflow/pull/7163#discussion_r381559962 ## File path: airflow/providers/cncf/kubernetes/operators/spark_kubernetes_operator.py ## @@ -0,0 +1,83 @@ +# +# Licensed to the Apache Software Foundation (ASF) under one +# or more contributor license agreements. See the NOTICE file +# distributed with this work for additional information +# regarding copyright ownership. The ASF licenses this file +# to you under the Apache License, Version 2.0 (the +# "License"); you may not use this file except in compliance +# with the License. You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, +# software distributed under the License is distributed on an +# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY +# KIND, either express or implied. See the License for the +# specific language governing permissions and limitations +# under the License. +from typing import Optional + +import yaml +from kubernetes import client + +from airflow.exceptions import AirflowException +from airflow.models import BaseOperator +from airflow.providers.cncf.kubernetes.hooks.kubernetes_hook import Kuberneteshook +from airflow.utils.decorators import apply_defaults + + +class SparkKubernetesOperator(BaseOperator): +""" +Creates sparkApplication object in kubernetes cluster: + .. seealso:: +For more detail about Spark Application Object have a look at the reference: + https://github.com/GoogleCloudPlatform/spark-on-k8s-operator/blob/master/docs/api-docs.md#sparkapplication + +:param sparkapplication_file: filepath to kubernetes custom_resource_definition of sparkApplication +:type sparkapplication_file: str +:param namespace: kubernetes namespace to put sparkApplication +:type namespace: str +:param conn_id: the connection to Kubernetes cluster +:type conn_id: str +""" + +template_fields = ['sparkapplication_file', 'namespace'] +template_ext = ('yaml', 'yml', 'json') +ui_color = '#f4a460' + +@apply_defaults +def __init__(self, + sparkapplication_file: str, + namespace: Optional[str] = None, + conn_id: str = 'kubernetes_default', + *args, **kwargs) -> None: +super().__init__(*args, **kwargs) +self.sparkapplication_file = sparkapplication_file +self.namespace = namespace +self.conn_id = conn_id + +def execute(self, context): +self.log.info("Creating sparkApplication") +hook = Kuberneteshook(conn_id=self.conn_id) +api_client = hook.get_conn() +api = client.CustomObjectsApi(api_client) Review comment: This feels like it should be in the hook, not the operator. This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[GitHub] [airflow] ashb commented on a change in pull request #7163: [AIRFLOW-6542] add spark-on-k8s operator/hook/sensor
ashb commented on a change in pull request #7163: [AIRFLOW-6542] add spark-on-k8s operator/hook/sensor URL: https://github.com/apache/airflow/pull/7163#discussion_r381559337 ## File path: airflow/providers/cncf/kubernetes/operators/spark_kubernetes_operator.py ## @@ -0,0 +1,83 @@ +# +# Licensed to the Apache Software Foundation (ASF) under one +# or more contributor license agreements. See the NOTICE file +# distributed with this work for additional information +# regarding copyright ownership. The ASF licenses this file +# to you under the Apache License, Version 2.0 (the +# "License"); you may not use this file except in compliance +# with the License. You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, +# software distributed under the License is distributed on an +# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY +# KIND, either express or implied. See the License for the +# specific language governing permissions and limitations +# under the License. +from typing import Optional + +import yaml +from kubernetes import client + +from airflow.exceptions import AirflowException +from airflow.models import BaseOperator +from airflow.providers.cncf.kubernetes.hooks.kubernetes_hook import Kuberneteshook +from airflow.utils.decorators import apply_defaults + + +class SparkKubernetesOperator(BaseOperator): +""" +Creates sparkApplication object in kubernetes cluster: + .. seealso:: +For more detail about Spark Application Object have a look at the reference: + https://github.com/GoogleCloudPlatform/spark-on-k8s-operator/blob/master/docs/api-docs.md#sparkapplication + +:param sparkapplication_file: filepath to kubernetes custom_resource_definition of sparkApplication +:type sparkapplication_file: str +:param namespace: kubernetes namespace to put sparkApplication +:type namespace: str +:param conn_id: the connection to Kubernetes cluster Review comment: ```suggestion :param kubernetes_conn_id: the connection to Kubernetes cluster ``` We prefix the connection id params in operators with the _type_ of the connection (which is done to make site-wide policies easier to implement) This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[GitHub] [airflow] ashb commented on a change in pull request #7163: [AIRFLOW-6542] add spark-on-k8s operator/hook/sensor
ashb commented on a change in pull request #7163: [AIRFLOW-6542] add spark-on-k8s operator/hook/sensor URL: https://github.com/apache/airflow/pull/7163#discussion_r381558940 ## File path: airflow/providers/cncf/kubernetes/operators/spark_kubernetes_operator.py ## @@ -0,0 +1,83 @@ +# +# Licensed to the Apache Software Foundation (ASF) under one +# or more contributor license agreements. See the NOTICE file +# distributed with this work for additional information +# regarding copyright ownership. The ASF licenses this file +# to you under the Apache License, Version 2.0 (the +# "License"); you may not use this file except in compliance +# with the License. You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, +# software distributed under the License is distributed on an +# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY +# KIND, either express or implied. See the License for the +# specific language governing permissions and limitations +# under the License. +from typing import Optional + +import yaml +from kubernetes import client + +from airflow.exceptions import AirflowException +from airflow.models import BaseOperator +from airflow.providers.cncf.kubernetes.hooks.kubernetes_hook import Kuberneteshook +from airflow.utils.decorators import apply_defaults + + +class SparkKubernetesOperator(BaseOperator): +""" +Creates sparkApplication object in kubernetes cluster: + .. seealso:: +For more detail about Spark Application Object have a look at the reference: + https://github.com/GoogleCloudPlatform/spark-on-k8s-operator/blob/master/docs/api-docs.md#sparkapplication + +:param sparkapplication_file: filepath to kubernetes custom_resource_definition of sparkApplication Review comment: ```suggestion :param application_file: filepath to kubernetes custom_resource_definition of sparkApplication ``` We're already in "spark" here, We don't have to mention spark again I don't think. This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[GitHub] [airflow] vsoch commented on issue #7191: [AIRFLOW-4030] second attempt to add singularity to airflow
vsoch commented on issue #7191: [AIRFLOW-4030] second attempt to add singularity to airflow URL: https://github.com/apache/airflow/pull/7191#issuecomment-588482545 Hmm, so I definitely added the disable line, but it doesn't seem to be taking: https://github.com/apache/airflow/pull/7191/commits/cef6c077927e996f5c910c872a775850a25596f0 It's triggering for line 65, so I suppose I can try adding it there... This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[GitHub] [airflow] ashb commented on a change in pull request #7163: [AIRFLOW-6542] add spark-on-k8s operator/hook/sensor
ashb commented on a change in pull request #7163: [AIRFLOW-6542] add spark-on-k8s operator/hook/sensor URL: https://github.com/apache/airflow/pull/7163#discussion_r381556793 ## File path: airflow/providers/cncf/kubernetes/hooks/kubernetes_hook.py ## @@ -0,0 +1,77 @@ +# Licensed to the Apache Software Foundation (ASF) under one +# or more contributor license agreements. See the NOTICE file +# distributed with this work for additional information +# regarding copyright ownership. The ASF licenses this file +# to you under the Apache License, Version 2.0 (the +# "License"); you may not use this file except in compliance +# with the License. You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, +# software distributed under the License is distributed on an +# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY +# KIND, either express or implied. See the License for the +# specific language governing permissions and limitations +# under the License. +import tempfile + +from kubernetes import client, config + +from airflow.hooks.base_hook import BaseHook + + +class Kuberneteshook(BaseHook): +""" +Creates Kubernetes API connection. + +:param conn_id: the connection to Kubernetes cluster +""" + +def __init__( +self, +conn_id="kubernetes_default" +): +self.connection = self.get_connection(conn_id) +self.extras = self.connection.extra_dejson + +def get_conn(self): +""" +Returns kubernetes api session for use with requests +""" + +if self._get_field(("in_cluster")): +self.log.debug("loading kube_config from: in_cluster configuration") +config.load_incluster_config() +elif self._get_field("kube_config") is None or self._get_field("kube_config") == '': +self.log.debug("loading kube_config from: default file") +else: +with tempfile.NamedTemporaryFile() as temp_config: +self.log.debug("loading kube_config from: connection kube_config") +temp_config.write(self._get_field("kube_config").encode()) +config.load_kube_config(temp_config.name) +temp_config.flush() +return client.ApiClient() + +def get_namespace(self): +""" +Returns the namespace that defined in the connection +""" +namespace = self._get_field("namespace") Review comment: ```suggestion return self._get_field("namespace", default="default") ``` would do for this method, no? This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[GitHub] [airflow] ashb commented on a change in pull request #7163: [AIRFLOW-6542] add spark-on-k8s operator/hook/sensor
ashb commented on a change in pull request #7163: [AIRFLOW-6542] add spark-on-k8s operator/hook/sensor URL: https://github.com/apache/airflow/pull/7163#discussion_r381556862 ## File path: airflow/providers/cncf/kubernetes/operators/spark_kubernetes_operator.py ## @@ -0,0 +1,83 @@ +# +# Licensed to the Apache Software Foundation (ASF) under one +# or more contributor license agreements. See the NOTICE file +# distributed with this work for additional information +# regarding copyright ownership. The ASF licenses this file +# to you under the Apache License, Version 2.0 (the +# "License"); you may not use this file except in compliance +# with the License. You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, +# software distributed under the License is distributed on an +# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY +# KIND, either express or implied. See the License for the +# specific language governing permissions and limitations +# under the License. +from typing import Optional + +import yaml +from kubernetes import client + +from airflow.exceptions import AirflowException +from airflow.models import BaseOperator +from airflow.providers.cncf.kubernetes.hooks.kubernetes_hook import Kuberneteshook +from airflow.utils.decorators import apply_defaults + + +class SparkKubernetesOperator(BaseOperator): +""" +Creates sparkApplication object in kubernetes cluster: + .. seealso:: Review comment: ```suggestion .. seealso:: ``` This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[GitHub] [airflow] mik-laj commented on issue #7458: [AIRFLOW-6838][WIP] Introduce real subcommands for Breeze
mik-laj commented on issue #7458: [AIRFLOW-6838][WIP] Introduce real subcommands for Breeze URL: https://github.com/apache/airflow/pull/7458#issuecomment-588480323 > I think we should get rid of the short one-letter options for sub-commands. Removed > ./breeze --help should return detailed explanation for that command only I agree, but I think this is an additional function that we can do independently. This will require much more code changes. In my opinion, we can divide it into two steps. First enter the commands and check users' opinions about them, and then refactor the code to further improve the code. This is a big change, because it will also require that we do not have one parser, but several smaller parsers. This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[GitHub] [airflow] ashb commented on a change in pull request #7163: [AIRFLOW-6542] add spark-on-k8s operator/hook/sensor
ashb commented on a change in pull request #7163: [AIRFLOW-6542] add spark-on-k8s operator/hook/sensor URL: https://github.com/apache/airflow/pull/7163#discussion_r381554309 ## File path: airflow/providers/cncf/kubernetes/hooks/kubernetes_hook.py ## @@ -0,0 +1,77 @@ +# Licensed to the Apache Software Foundation (ASF) under one +# or more contributor license agreements. See the NOTICE file +# distributed with this work for additional information +# regarding copyright ownership. The ASF licenses this file +# to you under the Apache License, Version 2.0 (the +# "License"); you may not use this file except in compliance +# with the License. You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, +# software distributed under the License is distributed on an +# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY +# KIND, either express or implied. See the License for the +# specific language governing permissions and limitations +# under the License. +import tempfile + +from kubernetes import client, config + +from airflow.hooks.base_hook import BaseHook + + +class Kuberneteshook(BaseHook): +""" +Creates Kubernetes API connection. + +:param conn_id: the connection to Kubernetes cluster +""" + +def __init__( +self, +conn_id="kubernetes_default" +): +self.connection = self.get_connection(conn_id) +self.extras = self.connection.extra_dejson + +def get_conn(self): +""" +Returns kubernetes api session for use with requests +""" + +if self._get_field(("in_cluster")): +self.log.debug("loading kube_config from: in_cluster configuration") +config.load_incluster_config() +elif self._get_field("kube_config") is None or self._get_field("kube_config") == '': +self.log.debug("loading kube_config from: default file") +else: +with tempfile.NamedTemporaryFile() as temp_config: +self.log.debug("loading kube_config from: connection kube_config") +temp_config.write(self._get_field("kube_config").encode()) +config.load_kube_config(temp_config.name) +temp_config.flush() Review comment: There's no point flushing the file _after_ it's been read by load_kube_config. Either the flush isn't needed, or it should be called before config.load_kube_config. Is there no option to load from a dict instead of having to write it to a temp file? This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[GitHub] [airflow] coopergillan commented on issue #7468: [AIRFLOW-6721] Organize hive tests
coopergillan commented on issue #7468: [AIRFLOW-6721] Organize hive tests URL: https://github.com/apache/airflow/pull/7468#issuecomment-588474051 @mik-laj - Done! https://issues.apache.org/jira/browse/AIRFLOW-6847 This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[GitHub] [airflow] kaxil commented on a change in pull request #7163: [AIRFLOW-6542] add spark-on-k8s operator/hook/sensor
kaxil commented on a change in pull request #7163: [AIRFLOW-6542] add spark-on-k8s operator/hook/sensor URL: https://github.com/apache/airflow/pull/7163#discussion_r381547632 ## File path: airflow/providers/cncf/kubernetes/hooks/kubernetes_hook.py ## @@ -0,0 +1,77 @@ +# Licensed to the Apache Software Foundation (ASF) under one +# or more contributor license agreements. See the NOTICE file +# distributed with this work for additional information +# regarding copyright ownership. The ASF licenses this file +# to you under the Apache License, Version 2.0 (the +# "License"); you may not use this file except in compliance +# with the License. You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, +# software distributed under the License is distributed on an +# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY +# KIND, either express or implied. See the License for the +# specific language governing permissions and limitations +# under the License. +import tempfile + +from kubernetes import client, config + +from airflow.hooks.base_hook import BaseHook + + +class Kuberneteshook(BaseHook): Review comment: ```suggestion class KubernetesHook(BaseHook): ``` This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[jira] [Created] (AIRFLOW-6847) Replace use of AIRFLOW_RUNALL_TESTS with pytest markers
Cooper Gillan created AIRFLOW-6847: -- Summary: Replace use of AIRFLOW_RUNALL_TESTS with pytest markers Key: AIRFLOW-6847 URL: https://issues.apache.org/jira/browse/AIRFLOW-6847 Project: Apache Airflow Issue Type: Improvement Components: tests Affects Versions: 1.10.9 Reporter: Cooper Gillan Building off work completed to split out Hive tests captured in AIRFLOW-6721 and initially worked on in https://github.com/apache/airflow/pull/7468, remove use of the {{AIRFLOW_RUNALL_TESTS}} environment variable for skipping certain tests in favor of using pytest markers. More info on airflow's use of pytest markers can be found here: https://github.com/apache/airflow/blob/master/TESTING.rst#airflow-integration-tests This ticket rose out of [this comment in PR #7468's review|https://github.com/apache/airflow/pull/7468#issuecomment-588458850]. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[GitHub] [airflow] kaxil commented on a change in pull request #7163: [AIRFLOW-6542] add spark-on-k8s operator/hook/sensor
kaxil commented on a change in pull request #7163: [AIRFLOW-6542] add spark-on-k8s operator/hook/sensor URL: https://github.com/apache/airflow/pull/7163#discussion_r381547572 ## File path: airflow/providers/cncf/kubernetes/hooks/kubernetes_hook.py ## @@ -0,0 +1,77 @@ +# Licensed to the Apache Software Foundation (ASF) under one +# or more contributor license agreements. See the NOTICE file Review comment: We should rename the file to `airflow/providers/cncf/kubernetes/hooks/kubernetes.py` like the other hooks This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[GitHub] [airflow] kaxil commented on a change in pull request #7163: [AIRFLOW-6542] add spark-on-k8s operator/hook/sensor
kaxil commented on a change in pull request #7163: [AIRFLOW-6542] add spark-on-k8s operator/hook/sensor URL: https://github.com/apache/airflow/pull/7163#discussion_r381547159 ## File path: airflow/models/connection.py ## @@ -61,6 +61,7 @@ "hiveserver2": ("airflow.providers.apache.hive.hooks.hive.HiveServer2Hook", "hiveserver2_conn_id"), "jdbc": ("airflow.providers.jdbc.hooks.jdbc.JdbcHook", "jdbc_conn_id"), "jira": ("airflow.providers.jira.hooks.jira.JiraHook", "jira_conn_id"), +"kubernetes": ("airflow.providers.cncf.kubernetes.hooks.kubernetes_hook", "kubernetes_conn_id"), Review comment: ```suggestion "kubernetes": ("airflow.providers.cncf.kubernetes.hooks.kubernetes.KubernetesHook", "kubernetes_conn_id"), ``` This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[GitHub] [airflow] mik-laj commented on a change in pull request #7458: [AIRFLOW-6838][WIP] Introduce real subcommands for Breeze
mik-laj commented on a change in pull request #7458: [AIRFLOW-6838][WIP] Introduce real subcommands for Breeze URL: https://github.com/apache/airflow/pull/7458#discussion_r381545067 ## File path: breeze-complete ## @@ -27,23 +27,47 @@ _BREEZE_DEFAULT_DOCKERHUB_USER="apache" _BREEZE_DEFAULT_DOCKERHUB_REPO="airflow" _BREEZE_SHORT_OPTIONS=" -h P: B: I: K Z X +h P: B: I: M: V: -s b O N -v y n C A +s N +v y n r p R L u -c D: H: e a -t: d: k x: S: F: +c D: H: " _BREEZE_LONG_OPTIONS=" -help python: backend: integration: start-kind-cluster recreate-kind-cluster stop-kind-cluster +help python: backend: integration: kubernetes-mode: kubernetes-version: -skip-mounting-local-sources build-only build-docs -verbose assume-yes assume-no toggle-suppress-cheatsheet toggle-suppress-asciiart +skip-mounting-local-sources +verbose assume-yes assume-no force-build-images force-pull-images force-clean-images use-local-cache push-images -cleanup-images dockerhub-user: dockerhub-repo: initialize-local-virtualenv setup-autocomplete -test-target: docker-compose: stop-environment execute-command: static-check: static-check-all-files: +cleanup-images dockerhub-user: dockerhub-repo: +" + +_BREEZE_COMMANDS=" +shell +O build-docs +b build-only +e initialize-local-virtualenv +a setup-autocomplete +k stop-environment +S static-check +F static-check-all-files +t test-target +h help +K start-kind-cluster +Z recreate-kind-cluster +X stop-kind-cluster +toggle-suppress-cheatsheet +toggle-suppress-asciiart +" + +_BREEZE_EXTRA_ARG_COMMANDS=" +S static-check Review comment: It's a mistake. Fixed. This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[GitHub] [airflow] mik-laj commented on issue #7458: [AIRFLOW-6838][WIP] Introduce real subcommands for Breeze
mik-laj commented on issue #7458: [AIRFLOW-6838][WIP] Introduce real subcommands for Breeze URL: https://github.com/apache/airflow/pull/7458#issuecomment-588470541 @potiuk I rebased. Can you look at it? This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[jira] [Updated] (AIRFLOW-6820) Split breeze into functions
[ https://issues.apache.org/jira/browse/AIRFLOW-6820?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jarek Potiuk updated AIRFLOW-6820: -- Fix Version/s: (was: 2.0.0) 1.10.10 > Split breeze into functions > --- > > Key: AIRFLOW-6820 > URL: https://issues.apache.org/jira/browse/AIRFLOW-6820 > Project: Apache Airflow > Issue Type: Improvement > Components: breeze >Affects Versions: 1.10.10 >Reporter: Jarek Potiuk >Priority: Major > Fix For: 1.10.10 > > -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Commented] (AIRFLOW-6839) Even more mypy speedup
[ https://issues.apache.org/jira/browse/AIRFLOW-6839?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17040438#comment-17040438 ] ASF subversion and git services commented on AIRFLOW-6839: -- Commit 28d5ade809b2510f83209129fea0078a3004946d in airflow's branch refs/heads/v1-10-test from Jarek Potiuk [ https://gitbox.apache.org/repos/asf?p=airflow.git;h=28d5ade ] [AIRFLOW-6839] even more mypy speed improvements (#7460) Require_serial:true is better choice than pass_filename: false as it can speed-up mypy for single file changes. Significant gains can be achieved for single file changes and no cache for all other files. This is majority of cases for our users who have pre-commits installed as hooks because most people change only few files and never run check with --all-files When just one file is changed and no cache is built, the difference is drastic: require_serial: true = 4s pass_filenames: false = 13s (cherry picked from commit 30898123ad5e5a2dbb55f83d859ef3b0e2f697d9) > Even more mypy speedup > -- > > Key: AIRFLOW-6839 > URL: https://issues.apache.org/jira/browse/AIRFLOW-6839 > Project: Apache Airflow > Issue Type: Improvement > Components: pre-commit >Affects Versions: 2.0.0, 1.10.9 >Reporter: Jarek Potiuk >Priority: Major > Fix For: 1.10.10 > > > require_serial:true is better choice than pass_filename: false as it can > speed-up mypy for single file changes. > Significant gains can be achieved for single file changes and no cache for > all other files. This is majority of cases for our users who have pre-commits > installed as hooks because most people change only few files and never run > check with --all-files > When just one file is changed and no cache is built, the difference is > drastic: > require_serial: true (4s) > > {code:java} > Run > mypy.PassedRun > mypy.Passed > real 0m4,080s user 0m0,283s sys 0m0,095s > {code} > pass_filenames: false: (13s) > {code:java} > time pre-commit run mypy time pre-commit run mypyRun > mypy.Passed > real 0m13,756suser 0m0,281ssys 0m0,099s{code} > -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Commented] (AIRFLOW-6841) Unbounded variable on Mac
[ https://issues.apache.org/jira/browse/AIRFLOW-6841?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17040437#comment-17040437 ] ASF subversion and git services commented on AIRFLOW-6841: -- Commit 4308bf443cfbfb4a52fa8d9c9a8a791487a5ad53 in airflow's branch refs/heads/v1-10-test from Jarek Potiuk [ https://gitbox.apache.org/repos/asf?p=airflow.git;h=4308bf4 ] [AIRFLOW-6841] Fixed unbounded variable on Mac (#7465) (cherry picked from commit ae69475522f8d5da4aa7e0cb3d7b107922c221de) > Unbounded variable on Mac > - > > Key: AIRFLOW-6841 > URL: https://issues.apache.org/jira/browse/AIRFLOW-6841 > Project: Apache Airflow > Issue Type: Bug > Components: breeze >Affects Versions: 2.0.0, 1.10.9 >Reporter: Jarek Potiuk >Priority: Major > > There is one more unbounded variable on Mac (old bash). > ./breeze: line 1198: @: unbound variable > > Finished the script breeze > Elapsed time spent in the script: 0 seconds > Exit code 0 > -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Resolved] (AIRFLOW-6841) Unbounded variable on Mac
[ https://issues.apache.org/jira/browse/AIRFLOW-6841?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jarek Potiuk resolved AIRFLOW-6841. --- Fix Version/s: 1.10.10 Resolution: Fixed > Unbounded variable on Mac > - > > Key: AIRFLOW-6841 > URL: https://issues.apache.org/jira/browse/AIRFLOW-6841 > Project: Apache Airflow > Issue Type: Bug > Components: breeze >Affects Versions: 2.0.0, 1.10.9 >Reporter: Jarek Potiuk >Priority: Major > Fix For: 1.10.10 > > > There is one more unbounded variable on Mac (old bash). > ./breeze: line 1198: @: unbound variable > > Finished the script breeze > Elapsed time spent in the script: 0 seconds > Exit code 0 > -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Commented] (AIRFLOW-6842) On Max checking for fix ownership returns space prefixed 0
[ https://issues.apache.org/jira/browse/AIRFLOW-6842?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17040439#comment-17040439 ] ASF subversion and git services commented on AIRFLOW-6842: -- Commit 6135f7dcd76db27acdb70e4394db7449fa1bcf7a in airflow's branch refs/heads/v1-10-test from Jarek Potiuk [ https://gitbox.apache.org/repos/asf?p=airflow.git;h=6135f7d ] [AIRFLOW-6842] Skip fixing ownership on Mac (#7469) (cherry picked from commit 1a9a9f7618f1c22e3e9a6ef4ec73b717c7760c7d) > On Max checking for fix ownership returns space prefixed 0 > -- > > Key: AIRFLOW-6842 > URL: https://issues.apache.org/jira/browse/AIRFLOW-6842 > Project: Apache Airflow > Issue Type: Bug > Components: breeze >Affects Versions: 2.0.0 >Reporter: Jarek Potiuk >Priority: Major > Fix For: 1.10.10 > > > On Mac fixing ownership takes time and it is basically never needed. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Commented] (AIRFLOW-6842) On Max checking for fix ownership returns space prefixed 0
[ https://issues.apache.org/jira/browse/AIRFLOW-6842?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17040435#comment-17040435 ] ASF subversion and git services commented on AIRFLOW-6842: -- Commit 1a9a9f7618f1c22e3e9a6ef4ec73b717c7760c7d in airflow's branch refs/heads/master from Jarek Potiuk [ https://gitbox.apache.org/repos/asf?p=airflow.git;h=1a9a9f7 ] [AIRFLOW-6842] Skip fixing ownership on Mac (#7469) > On Max checking for fix ownership returns space prefixed 0 > -- > > Key: AIRFLOW-6842 > URL: https://issues.apache.org/jira/browse/AIRFLOW-6842 > Project: Apache Airflow > Issue Type: Bug > Components: breeze >Affects Versions: 2.0.0 >Reporter: Jarek Potiuk >Priority: Major > Fix For: 1.10.10 > > > On Mac fixing ownership takes time and it is basically never needed. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[GitHub] [airflow] vsoch commented on issue #7191: [AIRFLOW-4030] second attempt to add singularity to airflow
vsoch commented on issue #7191: [AIRFLOW-4030] second attempt to add singularity to airflow URL: https://github.com/apache/airflow/pull/7191#issuecomment-588463321 @potiuk I tried this huge build multiple times before, and they all resulted in errors. I'm glad to try it again, but I've had much better luck with reading the errors and addressing the errors verbatim. This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[jira] [Commented] (AIRFLOW-6721) Reorganize and clean up Hive tests
[ https://issues.apache.org/jira/browse/AIRFLOW-6721?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17040431#comment-17040431 ] ASF subversion and git services commented on AIRFLOW-6721: -- Commit e91d615e02758da6b9f3cf6be502f0114d866ea3 in airflow's branch refs/heads/master from Cooper Gillan [ https://gitbox.apache.org/repos/asf?p=airflow.git;h=e91d615 ] [AIRFLOW-6721] Organize Apache Hive tests (#7468) * [AIRFLOW-6721] Move WebHdfsSensor tests to own module Move the WebHdfsSensor tests out of the operator tests and into a separate module. * [AIRFLOW-6721] Move TestHiveEnvironment to test module Movee the TestHiveEnvironment class and setUp method out to the hive test module so that it can be used in other hive tests. * [AIRFLOW-6721] Move Hive Stats tests to separate module Move skipped test for HiveStatsCollectionOperator to the test_hive_stats module to make it easier to locate. * [AIRFLOW-6721] Move NamedHivePartitionSensor tests Move the NamedHivePartitionSensor tests from the operators module into the test module for NamedHivePartitionSensor. * [AIRFLOW-6721] Move HivePartitionSensor tests Move the HivePartitionSensor tests from the operators module into the a new test module. * [AIRFLOW-6721] Move HiveToMySqlTransfer test Move the HiveToMySqlTransfer test in the operators module into the existing test module and use the shared test environment for these tests. * [AIRFLOW-6721] Move HdfsSensor test to separate module Move the HdfsSensor test in the operators module into a new test module and use the shared test environment. * [AIRFLOW-6721] Move Hive2SambaOperator test Move the Hive2SambaOperator test in the operators module into the existing test module and use the shared test environment. * [AIRFLOW-6721] Move Hive MetastorePartitionSensor test Move the MetastorePartitionSensor test from the operators module into a new module and use the shared test environment. * [AIRFLOW-6721] Move Hive PrestoToMySqlTransfer test Move the PrestoToMySqlTransfer test from the operators module into an existing test module and use shared Hive test environment. * [AIRFLOW-6721] Move PrestoCheckOperator test Move the PrestoCheckOperator test from the Hive operators module into a new Presto operators test module while still using the shared Hive test environment. * [AIRFLOW-6721] Move Hive SqlSensor test Move the SqlSensor in the Hive operators module into the existing SqlSensor test module and label it as such. Use the shared test environment for all tests, ensuring that original setUp properties are preserved for the existing tests. > Reorganize and clean up Hive tests > -- > > Key: AIRFLOW-6721 > URL: https://issues.apache.org/jira/browse/AIRFLOW-6721 > Project: Apache Airflow > Issue Type: Sub-task > Components: tests >Affects Versions: 1.10.7 >Reporter: Cooper Gillan >Assignee: Cooper Gillan >Priority: Major > Fix For: 2.0.0 > > > From [this PR > comment|https://github.com/apache/airflow/pull/7316#issuecomment-580946162]: > https://github.com/apache/airflow/blob/master/tests/providers/apache/hive/operators/test_hive.py#L29-L40 > This file contains tests for many unrelated components. I think it would be > useful to transfer some tests to other files. > Here is an example: PolideaInternal@a91b612 > To avoid conflicts you can do it in one PR, but it would be useful for each > move to be in a separate commit. This will facilitate the review. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Commented] (AIRFLOW-6721) Reorganize and clean up Hive tests
[ https://issues.apache.org/jira/browse/AIRFLOW-6721?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17040429#comment-17040429 ] ASF subversion and git services commented on AIRFLOW-6721: -- Commit e91d615e02758da6b9f3cf6be502f0114d866ea3 in airflow's branch refs/heads/master from Cooper Gillan [ https://gitbox.apache.org/repos/asf?p=airflow.git;h=e91d615 ] [AIRFLOW-6721] Organize Apache Hive tests (#7468) * [AIRFLOW-6721] Move WebHdfsSensor tests to own module Move the WebHdfsSensor tests out of the operator tests and into a separate module. * [AIRFLOW-6721] Move TestHiveEnvironment to test module Movee the TestHiveEnvironment class and setUp method out to the hive test module so that it can be used in other hive tests. * [AIRFLOW-6721] Move Hive Stats tests to separate module Move skipped test for HiveStatsCollectionOperator to the test_hive_stats module to make it easier to locate. * [AIRFLOW-6721] Move NamedHivePartitionSensor tests Move the NamedHivePartitionSensor tests from the operators module into the test module for NamedHivePartitionSensor. * [AIRFLOW-6721] Move HivePartitionSensor tests Move the HivePartitionSensor tests from the operators module into the a new test module. * [AIRFLOW-6721] Move HiveToMySqlTransfer test Move the HiveToMySqlTransfer test in the operators module into the existing test module and use the shared test environment for these tests. * [AIRFLOW-6721] Move HdfsSensor test to separate module Move the HdfsSensor test in the operators module into a new test module and use the shared test environment. * [AIRFLOW-6721] Move Hive2SambaOperator test Move the Hive2SambaOperator test in the operators module into the existing test module and use the shared test environment. * [AIRFLOW-6721] Move Hive MetastorePartitionSensor test Move the MetastorePartitionSensor test from the operators module into a new module and use the shared test environment. * [AIRFLOW-6721] Move Hive PrestoToMySqlTransfer test Move the PrestoToMySqlTransfer test from the operators module into an existing test module and use shared Hive test environment. * [AIRFLOW-6721] Move PrestoCheckOperator test Move the PrestoCheckOperator test from the Hive operators module into a new Presto operators test module while still using the shared Hive test environment. * [AIRFLOW-6721] Move Hive SqlSensor test Move the SqlSensor in the Hive operators module into the existing SqlSensor test module and label it as such. Use the shared test environment for all tests, ensuring that original setUp properties are preserved for the existing tests. > Reorganize and clean up Hive tests > -- > > Key: AIRFLOW-6721 > URL: https://issues.apache.org/jira/browse/AIRFLOW-6721 > Project: Apache Airflow > Issue Type: Sub-task > Components: tests >Affects Versions: 1.10.7 >Reporter: Cooper Gillan >Assignee: Cooper Gillan >Priority: Major > Fix For: 2.0.0 > > > From [this PR > comment|https://github.com/apache/airflow/pull/7316#issuecomment-580946162]: > https://github.com/apache/airflow/blob/master/tests/providers/apache/hive/operators/test_hive.py#L29-L40 > This file contains tests for many unrelated components. I think it would be > useful to transfer some tests to other files. > Here is an example: PolideaInternal@a91b612 > To avoid conflicts you can do it in one PR, but it would be useful for each > move to be in a separate commit. This will facilitate the review. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Commented] (AIRFLOW-6721) Reorganize and clean up Hive tests
[ https://issues.apache.org/jira/browse/AIRFLOW-6721?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17040419#comment-17040419 ] ASF subversion and git services commented on AIRFLOW-6721: -- Commit e91d615e02758da6b9f3cf6be502f0114d866ea3 in airflow's branch refs/heads/master from Cooper Gillan [ https://gitbox.apache.org/repos/asf?p=airflow.git;h=e91d615 ] [AIRFLOW-6721] Organize Apache Hive tests (#7468) * [AIRFLOW-6721] Move WebHdfsSensor tests to own module Move the WebHdfsSensor tests out of the operator tests and into a separate module. * [AIRFLOW-6721] Move TestHiveEnvironment to test module Movee the TestHiveEnvironment class and setUp method out to the hive test module so that it can be used in other hive tests. * [AIRFLOW-6721] Move Hive Stats tests to separate module Move skipped test for HiveStatsCollectionOperator to the test_hive_stats module to make it easier to locate. * [AIRFLOW-6721] Move NamedHivePartitionSensor tests Move the NamedHivePartitionSensor tests from the operators module into the test module for NamedHivePartitionSensor. * [AIRFLOW-6721] Move HivePartitionSensor tests Move the HivePartitionSensor tests from the operators module into the a new test module. * [AIRFLOW-6721] Move HiveToMySqlTransfer test Move the HiveToMySqlTransfer test in the operators module into the existing test module and use the shared test environment for these tests. * [AIRFLOW-6721] Move HdfsSensor test to separate module Move the HdfsSensor test in the operators module into a new test module and use the shared test environment. * [AIRFLOW-6721] Move Hive2SambaOperator test Move the Hive2SambaOperator test in the operators module into the existing test module and use the shared test environment. * [AIRFLOW-6721] Move Hive MetastorePartitionSensor test Move the MetastorePartitionSensor test from the operators module into a new module and use the shared test environment. * [AIRFLOW-6721] Move Hive PrestoToMySqlTransfer test Move the PrestoToMySqlTransfer test from the operators module into an existing test module and use shared Hive test environment. * [AIRFLOW-6721] Move PrestoCheckOperator test Move the PrestoCheckOperator test from the Hive operators module into a new Presto operators test module while still using the shared Hive test environment. * [AIRFLOW-6721] Move Hive SqlSensor test Move the SqlSensor in the Hive operators module into the existing SqlSensor test module and label it as such. Use the shared test environment for all tests, ensuring that original setUp properties are preserved for the existing tests. > Reorganize and clean up Hive tests > -- > > Key: AIRFLOW-6721 > URL: https://issues.apache.org/jira/browse/AIRFLOW-6721 > Project: Apache Airflow > Issue Type: Sub-task > Components: tests >Affects Versions: 1.10.7 >Reporter: Cooper Gillan >Assignee: Cooper Gillan >Priority: Major > Fix For: 2.0.0 > > > From [this PR > comment|https://github.com/apache/airflow/pull/7316#issuecomment-580946162]: > https://github.com/apache/airflow/blob/master/tests/providers/apache/hive/operators/test_hive.py#L29-L40 > This file contains tests for many unrelated components. I think it would be > useful to transfer some tests to other files. > Here is an example: PolideaInternal@a91b612 > To avoid conflicts you can do it in one PR, but it would be useful for each > move to be in a separate commit. This will facilitate the review. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Commented] (AIRFLOW-6721) Reorganize and clean up Hive tests
[ https://issues.apache.org/jira/browse/AIRFLOW-6721?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17040425#comment-17040425 ] ASF subversion and git services commented on AIRFLOW-6721: -- Commit e91d615e02758da6b9f3cf6be502f0114d866ea3 in airflow's branch refs/heads/master from Cooper Gillan [ https://gitbox.apache.org/repos/asf?p=airflow.git;h=e91d615 ] [AIRFLOW-6721] Organize Apache Hive tests (#7468) * [AIRFLOW-6721] Move WebHdfsSensor tests to own module Move the WebHdfsSensor tests out of the operator tests and into a separate module. * [AIRFLOW-6721] Move TestHiveEnvironment to test module Movee the TestHiveEnvironment class and setUp method out to the hive test module so that it can be used in other hive tests. * [AIRFLOW-6721] Move Hive Stats tests to separate module Move skipped test for HiveStatsCollectionOperator to the test_hive_stats module to make it easier to locate. * [AIRFLOW-6721] Move NamedHivePartitionSensor tests Move the NamedHivePartitionSensor tests from the operators module into the test module for NamedHivePartitionSensor. * [AIRFLOW-6721] Move HivePartitionSensor tests Move the HivePartitionSensor tests from the operators module into the a new test module. * [AIRFLOW-6721] Move HiveToMySqlTransfer test Move the HiveToMySqlTransfer test in the operators module into the existing test module and use the shared test environment for these tests. * [AIRFLOW-6721] Move HdfsSensor test to separate module Move the HdfsSensor test in the operators module into a new test module and use the shared test environment. * [AIRFLOW-6721] Move Hive2SambaOperator test Move the Hive2SambaOperator test in the operators module into the existing test module and use the shared test environment. * [AIRFLOW-6721] Move Hive MetastorePartitionSensor test Move the MetastorePartitionSensor test from the operators module into a new module and use the shared test environment. * [AIRFLOW-6721] Move Hive PrestoToMySqlTransfer test Move the PrestoToMySqlTransfer test from the operators module into an existing test module and use shared Hive test environment. * [AIRFLOW-6721] Move PrestoCheckOperator test Move the PrestoCheckOperator test from the Hive operators module into a new Presto operators test module while still using the shared Hive test environment. * [AIRFLOW-6721] Move Hive SqlSensor test Move the SqlSensor in the Hive operators module into the existing SqlSensor test module and label it as such. Use the shared test environment for all tests, ensuring that original setUp properties are preserved for the existing tests. > Reorganize and clean up Hive tests > -- > > Key: AIRFLOW-6721 > URL: https://issues.apache.org/jira/browse/AIRFLOW-6721 > Project: Apache Airflow > Issue Type: Sub-task > Components: tests >Affects Versions: 1.10.7 >Reporter: Cooper Gillan >Assignee: Cooper Gillan >Priority: Major > Fix For: 2.0.0 > > > From [this PR > comment|https://github.com/apache/airflow/pull/7316#issuecomment-580946162]: > https://github.com/apache/airflow/blob/master/tests/providers/apache/hive/operators/test_hive.py#L29-L40 > This file contains tests for many unrelated components. I think it would be > useful to transfer some tests to other files. > Here is an example: PolideaInternal@a91b612 > To avoid conflicts you can do it in one PR, but it would be useful for each > move to be in a separate commit. This will facilitate the review. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Commented] (AIRFLOW-6721) Reorganize and clean up Hive tests
[ https://issues.apache.org/jira/browse/AIRFLOW-6721?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17040423#comment-17040423 ] ASF subversion and git services commented on AIRFLOW-6721: -- Commit e91d615e02758da6b9f3cf6be502f0114d866ea3 in airflow's branch refs/heads/master from Cooper Gillan [ https://gitbox.apache.org/repos/asf?p=airflow.git;h=e91d615 ] [AIRFLOW-6721] Organize Apache Hive tests (#7468) * [AIRFLOW-6721] Move WebHdfsSensor tests to own module Move the WebHdfsSensor tests out of the operator tests and into a separate module. * [AIRFLOW-6721] Move TestHiveEnvironment to test module Movee the TestHiveEnvironment class and setUp method out to the hive test module so that it can be used in other hive tests. * [AIRFLOW-6721] Move Hive Stats tests to separate module Move skipped test for HiveStatsCollectionOperator to the test_hive_stats module to make it easier to locate. * [AIRFLOW-6721] Move NamedHivePartitionSensor tests Move the NamedHivePartitionSensor tests from the operators module into the test module for NamedHivePartitionSensor. * [AIRFLOW-6721] Move HivePartitionSensor tests Move the HivePartitionSensor tests from the operators module into the a new test module. * [AIRFLOW-6721] Move HiveToMySqlTransfer test Move the HiveToMySqlTransfer test in the operators module into the existing test module and use the shared test environment for these tests. * [AIRFLOW-6721] Move HdfsSensor test to separate module Move the HdfsSensor test in the operators module into a new test module and use the shared test environment. * [AIRFLOW-6721] Move Hive2SambaOperator test Move the Hive2SambaOperator test in the operators module into the existing test module and use the shared test environment. * [AIRFLOW-6721] Move Hive MetastorePartitionSensor test Move the MetastorePartitionSensor test from the operators module into a new module and use the shared test environment. * [AIRFLOW-6721] Move Hive PrestoToMySqlTransfer test Move the PrestoToMySqlTransfer test from the operators module into an existing test module and use shared Hive test environment. * [AIRFLOW-6721] Move PrestoCheckOperator test Move the PrestoCheckOperator test from the Hive operators module into a new Presto operators test module while still using the shared Hive test environment. * [AIRFLOW-6721] Move Hive SqlSensor test Move the SqlSensor in the Hive operators module into the existing SqlSensor test module and label it as such. Use the shared test environment for all tests, ensuring that original setUp properties are preserved for the existing tests. > Reorganize and clean up Hive tests > -- > > Key: AIRFLOW-6721 > URL: https://issues.apache.org/jira/browse/AIRFLOW-6721 > Project: Apache Airflow > Issue Type: Sub-task > Components: tests >Affects Versions: 1.10.7 >Reporter: Cooper Gillan >Assignee: Cooper Gillan >Priority: Major > Fix For: 2.0.0 > > > From [this PR > comment|https://github.com/apache/airflow/pull/7316#issuecomment-580946162]: > https://github.com/apache/airflow/blob/master/tests/providers/apache/hive/operators/test_hive.py#L29-L40 > This file contains tests for many unrelated components. I think it would be > useful to transfer some tests to other files. > Here is an example: PolideaInternal@a91b612 > To avoid conflicts you can do it in one PR, but it would be useful for each > move to be in a separate commit. This will facilitate the review. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Commented] (AIRFLOW-6721) Reorganize and clean up Hive tests
[ https://issues.apache.org/jira/browse/AIRFLOW-6721?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17040426#comment-17040426 ] ASF subversion and git services commented on AIRFLOW-6721: -- Commit e91d615e02758da6b9f3cf6be502f0114d866ea3 in airflow's branch refs/heads/master from Cooper Gillan [ https://gitbox.apache.org/repos/asf?p=airflow.git;h=e91d615 ] [AIRFLOW-6721] Organize Apache Hive tests (#7468) * [AIRFLOW-6721] Move WebHdfsSensor tests to own module Move the WebHdfsSensor tests out of the operator tests and into a separate module. * [AIRFLOW-6721] Move TestHiveEnvironment to test module Movee the TestHiveEnvironment class and setUp method out to the hive test module so that it can be used in other hive tests. * [AIRFLOW-6721] Move Hive Stats tests to separate module Move skipped test for HiveStatsCollectionOperator to the test_hive_stats module to make it easier to locate. * [AIRFLOW-6721] Move NamedHivePartitionSensor tests Move the NamedHivePartitionSensor tests from the operators module into the test module for NamedHivePartitionSensor. * [AIRFLOW-6721] Move HivePartitionSensor tests Move the HivePartitionSensor tests from the operators module into the a new test module. * [AIRFLOW-6721] Move HiveToMySqlTransfer test Move the HiveToMySqlTransfer test in the operators module into the existing test module and use the shared test environment for these tests. * [AIRFLOW-6721] Move HdfsSensor test to separate module Move the HdfsSensor test in the operators module into a new test module and use the shared test environment. * [AIRFLOW-6721] Move Hive2SambaOperator test Move the Hive2SambaOperator test in the operators module into the existing test module and use the shared test environment. * [AIRFLOW-6721] Move Hive MetastorePartitionSensor test Move the MetastorePartitionSensor test from the operators module into a new module and use the shared test environment. * [AIRFLOW-6721] Move Hive PrestoToMySqlTransfer test Move the PrestoToMySqlTransfer test from the operators module into an existing test module and use shared Hive test environment. * [AIRFLOW-6721] Move PrestoCheckOperator test Move the PrestoCheckOperator test from the Hive operators module into a new Presto operators test module while still using the shared Hive test environment. * [AIRFLOW-6721] Move Hive SqlSensor test Move the SqlSensor in the Hive operators module into the existing SqlSensor test module and label it as such. Use the shared test environment for all tests, ensuring that original setUp properties are preserved for the existing tests. > Reorganize and clean up Hive tests > -- > > Key: AIRFLOW-6721 > URL: https://issues.apache.org/jira/browse/AIRFLOW-6721 > Project: Apache Airflow > Issue Type: Sub-task > Components: tests >Affects Versions: 1.10.7 >Reporter: Cooper Gillan >Assignee: Cooper Gillan >Priority: Major > Fix For: 2.0.0 > > > From [this PR > comment|https://github.com/apache/airflow/pull/7316#issuecomment-580946162]: > https://github.com/apache/airflow/blob/master/tests/providers/apache/hive/operators/test_hive.py#L29-L40 > This file contains tests for many unrelated components. I think it would be > useful to transfer some tests to other files. > Here is an example: PolideaInternal@a91b612 > To avoid conflicts you can do it in one PR, but it would be useful for each > move to be in a separate commit. This will facilitate the review. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Commented] (AIRFLOW-6721) Reorganize and clean up Hive tests
[ https://issues.apache.org/jira/browse/AIRFLOW-6721?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17040421#comment-17040421 ] ASF subversion and git services commented on AIRFLOW-6721: -- Commit e91d615e02758da6b9f3cf6be502f0114d866ea3 in airflow's branch refs/heads/master from Cooper Gillan [ https://gitbox.apache.org/repos/asf?p=airflow.git;h=e91d615 ] [AIRFLOW-6721] Organize Apache Hive tests (#7468) * [AIRFLOW-6721] Move WebHdfsSensor tests to own module Move the WebHdfsSensor tests out of the operator tests and into a separate module. * [AIRFLOW-6721] Move TestHiveEnvironment to test module Movee the TestHiveEnvironment class and setUp method out to the hive test module so that it can be used in other hive tests. * [AIRFLOW-6721] Move Hive Stats tests to separate module Move skipped test for HiveStatsCollectionOperator to the test_hive_stats module to make it easier to locate. * [AIRFLOW-6721] Move NamedHivePartitionSensor tests Move the NamedHivePartitionSensor tests from the operators module into the test module for NamedHivePartitionSensor. * [AIRFLOW-6721] Move HivePartitionSensor tests Move the HivePartitionSensor tests from the operators module into the a new test module. * [AIRFLOW-6721] Move HiveToMySqlTransfer test Move the HiveToMySqlTransfer test in the operators module into the existing test module and use the shared test environment for these tests. * [AIRFLOW-6721] Move HdfsSensor test to separate module Move the HdfsSensor test in the operators module into a new test module and use the shared test environment. * [AIRFLOW-6721] Move Hive2SambaOperator test Move the Hive2SambaOperator test in the operators module into the existing test module and use the shared test environment. * [AIRFLOW-6721] Move Hive MetastorePartitionSensor test Move the MetastorePartitionSensor test from the operators module into a new module and use the shared test environment. * [AIRFLOW-6721] Move Hive PrestoToMySqlTransfer test Move the PrestoToMySqlTransfer test from the operators module into an existing test module and use shared Hive test environment. * [AIRFLOW-6721] Move PrestoCheckOperator test Move the PrestoCheckOperator test from the Hive operators module into a new Presto operators test module while still using the shared Hive test environment. * [AIRFLOW-6721] Move Hive SqlSensor test Move the SqlSensor in the Hive operators module into the existing SqlSensor test module and label it as such. Use the shared test environment for all tests, ensuring that original setUp properties are preserved for the existing tests. > Reorganize and clean up Hive tests > -- > > Key: AIRFLOW-6721 > URL: https://issues.apache.org/jira/browse/AIRFLOW-6721 > Project: Apache Airflow > Issue Type: Sub-task > Components: tests >Affects Versions: 1.10.7 >Reporter: Cooper Gillan >Assignee: Cooper Gillan >Priority: Major > Fix For: 2.0.0 > > > From [this PR > comment|https://github.com/apache/airflow/pull/7316#issuecomment-580946162]: > https://github.com/apache/airflow/blob/master/tests/providers/apache/hive/operators/test_hive.py#L29-L40 > This file contains tests for many unrelated components. I think it would be > useful to transfer some tests to other files. > Here is an example: PolideaInternal@a91b612 > To avoid conflicts you can do it in one PR, but it would be useful for each > move to be in a separate commit. This will facilitate the review. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Commented] (AIRFLOW-6721) Reorganize and clean up Hive tests
[ https://issues.apache.org/jira/browse/AIRFLOW-6721?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17040422#comment-17040422 ] ASF subversion and git services commented on AIRFLOW-6721: -- Commit e91d615e02758da6b9f3cf6be502f0114d866ea3 in airflow's branch refs/heads/master from Cooper Gillan [ https://gitbox.apache.org/repos/asf?p=airflow.git;h=e91d615 ] [AIRFLOW-6721] Organize Apache Hive tests (#7468) * [AIRFLOW-6721] Move WebHdfsSensor tests to own module Move the WebHdfsSensor tests out of the operator tests and into a separate module. * [AIRFLOW-6721] Move TestHiveEnvironment to test module Movee the TestHiveEnvironment class and setUp method out to the hive test module so that it can be used in other hive tests. * [AIRFLOW-6721] Move Hive Stats tests to separate module Move skipped test for HiveStatsCollectionOperator to the test_hive_stats module to make it easier to locate. * [AIRFLOW-6721] Move NamedHivePartitionSensor tests Move the NamedHivePartitionSensor tests from the operators module into the test module for NamedHivePartitionSensor. * [AIRFLOW-6721] Move HivePartitionSensor tests Move the HivePartitionSensor tests from the operators module into the a new test module. * [AIRFLOW-6721] Move HiveToMySqlTransfer test Move the HiveToMySqlTransfer test in the operators module into the existing test module and use the shared test environment for these tests. * [AIRFLOW-6721] Move HdfsSensor test to separate module Move the HdfsSensor test in the operators module into a new test module and use the shared test environment. * [AIRFLOW-6721] Move Hive2SambaOperator test Move the Hive2SambaOperator test in the operators module into the existing test module and use the shared test environment. * [AIRFLOW-6721] Move Hive MetastorePartitionSensor test Move the MetastorePartitionSensor test from the operators module into a new module and use the shared test environment. * [AIRFLOW-6721] Move Hive PrestoToMySqlTransfer test Move the PrestoToMySqlTransfer test from the operators module into an existing test module and use shared Hive test environment. * [AIRFLOW-6721] Move PrestoCheckOperator test Move the PrestoCheckOperator test from the Hive operators module into a new Presto operators test module while still using the shared Hive test environment. * [AIRFLOW-6721] Move Hive SqlSensor test Move the SqlSensor in the Hive operators module into the existing SqlSensor test module and label it as such. Use the shared test environment for all tests, ensuring that original setUp properties are preserved for the existing tests. > Reorganize and clean up Hive tests > -- > > Key: AIRFLOW-6721 > URL: https://issues.apache.org/jira/browse/AIRFLOW-6721 > Project: Apache Airflow > Issue Type: Sub-task > Components: tests >Affects Versions: 1.10.7 >Reporter: Cooper Gillan >Assignee: Cooper Gillan >Priority: Major > Fix For: 2.0.0 > > > From [this PR > comment|https://github.com/apache/airflow/pull/7316#issuecomment-580946162]: > https://github.com/apache/airflow/blob/master/tests/providers/apache/hive/operators/test_hive.py#L29-L40 > This file contains tests for many unrelated components. I think it would be > useful to transfer some tests to other files. > Here is an example: PolideaInternal@a91b612 > To avoid conflicts you can do it in one PR, but it would be useful for each > move to be in a separate commit. This will facilitate the review. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Commented] (AIRFLOW-6721) Reorganize and clean up Hive tests
[ https://issues.apache.org/jira/browse/AIRFLOW-6721?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17040430#comment-17040430 ] ASF subversion and git services commented on AIRFLOW-6721: -- Commit e91d615e02758da6b9f3cf6be502f0114d866ea3 in airflow's branch refs/heads/master from Cooper Gillan [ https://gitbox.apache.org/repos/asf?p=airflow.git;h=e91d615 ] [AIRFLOW-6721] Organize Apache Hive tests (#7468) * [AIRFLOW-6721] Move WebHdfsSensor tests to own module Move the WebHdfsSensor tests out of the operator tests and into a separate module. * [AIRFLOW-6721] Move TestHiveEnvironment to test module Movee the TestHiveEnvironment class and setUp method out to the hive test module so that it can be used in other hive tests. * [AIRFLOW-6721] Move Hive Stats tests to separate module Move skipped test for HiveStatsCollectionOperator to the test_hive_stats module to make it easier to locate. * [AIRFLOW-6721] Move NamedHivePartitionSensor tests Move the NamedHivePartitionSensor tests from the operators module into the test module for NamedHivePartitionSensor. * [AIRFLOW-6721] Move HivePartitionSensor tests Move the HivePartitionSensor tests from the operators module into the a new test module. * [AIRFLOW-6721] Move HiveToMySqlTransfer test Move the HiveToMySqlTransfer test in the operators module into the existing test module and use the shared test environment for these tests. * [AIRFLOW-6721] Move HdfsSensor test to separate module Move the HdfsSensor test in the operators module into a new test module and use the shared test environment. * [AIRFLOW-6721] Move Hive2SambaOperator test Move the Hive2SambaOperator test in the operators module into the existing test module and use the shared test environment. * [AIRFLOW-6721] Move Hive MetastorePartitionSensor test Move the MetastorePartitionSensor test from the operators module into a new module and use the shared test environment. * [AIRFLOW-6721] Move Hive PrestoToMySqlTransfer test Move the PrestoToMySqlTransfer test from the operators module into an existing test module and use shared Hive test environment. * [AIRFLOW-6721] Move PrestoCheckOperator test Move the PrestoCheckOperator test from the Hive operators module into a new Presto operators test module while still using the shared Hive test environment. * [AIRFLOW-6721] Move Hive SqlSensor test Move the SqlSensor in the Hive operators module into the existing SqlSensor test module and label it as such. Use the shared test environment for all tests, ensuring that original setUp properties are preserved for the existing tests. > Reorganize and clean up Hive tests > -- > > Key: AIRFLOW-6721 > URL: https://issues.apache.org/jira/browse/AIRFLOW-6721 > Project: Apache Airflow > Issue Type: Sub-task > Components: tests >Affects Versions: 1.10.7 >Reporter: Cooper Gillan >Assignee: Cooper Gillan >Priority: Major > Fix For: 2.0.0 > > > From [this PR > comment|https://github.com/apache/airflow/pull/7316#issuecomment-580946162]: > https://github.com/apache/airflow/blob/master/tests/providers/apache/hive/operators/test_hive.py#L29-L40 > This file contains tests for many unrelated components. I think it would be > useful to transfer some tests to other files. > Here is an example: PolideaInternal@a91b612 > To avoid conflicts you can do it in one PR, but it would be useful for each > move to be in a separate commit. This will facilitate the review. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Commented] (AIRFLOW-6721) Reorganize and clean up Hive tests
[ https://issues.apache.org/jira/browse/AIRFLOW-6721?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17040420#comment-17040420 ] ASF subversion and git services commented on AIRFLOW-6721: -- Commit e91d615e02758da6b9f3cf6be502f0114d866ea3 in airflow's branch refs/heads/master from Cooper Gillan [ https://gitbox.apache.org/repos/asf?p=airflow.git;h=e91d615 ] [AIRFLOW-6721] Organize Apache Hive tests (#7468) * [AIRFLOW-6721] Move WebHdfsSensor tests to own module Move the WebHdfsSensor tests out of the operator tests and into a separate module. * [AIRFLOW-6721] Move TestHiveEnvironment to test module Movee the TestHiveEnvironment class and setUp method out to the hive test module so that it can be used in other hive tests. * [AIRFLOW-6721] Move Hive Stats tests to separate module Move skipped test for HiveStatsCollectionOperator to the test_hive_stats module to make it easier to locate. * [AIRFLOW-6721] Move NamedHivePartitionSensor tests Move the NamedHivePartitionSensor tests from the operators module into the test module for NamedHivePartitionSensor. * [AIRFLOW-6721] Move HivePartitionSensor tests Move the HivePartitionSensor tests from the operators module into the a new test module. * [AIRFLOW-6721] Move HiveToMySqlTransfer test Move the HiveToMySqlTransfer test in the operators module into the existing test module and use the shared test environment for these tests. * [AIRFLOW-6721] Move HdfsSensor test to separate module Move the HdfsSensor test in the operators module into a new test module and use the shared test environment. * [AIRFLOW-6721] Move Hive2SambaOperator test Move the Hive2SambaOperator test in the operators module into the existing test module and use the shared test environment. * [AIRFLOW-6721] Move Hive MetastorePartitionSensor test Move the MetastorePartitionSensor test from the operators module into a new module and use the shared test environment. * [AIRFLOW-6721] Move Hive PrestoToMySqlTransfer test Move the PrestoToMySqlTransfer test from the operators module into an existing test module and use shared Hive test environment. * [AIRFLOW-6721] Move PrestoCheckOperator test Move the PrestoCheckOperator test from the Hive operators module into a new Presto operators test module while still using the shared Hive test environment. * [AIRFLOW-6721] Move Hive SqlSensor test Move the SqlSensor in the Hive operators module into the existing SqlSensor test module and label it as such. Use the shared test environment for all tests, ensuring that original setUp properties are preserved for the existing tests. > Reorganize and clean up Hive tests > -- > > Key: AIRFLOW-6721 > URL: https://issues.apache.org/jira/browse/AIRFLOW-6721 > Project: Apache Airflow > Issue Type: Sub-task > Components: tests >Affects Versions: 1.10.7 >Reporter: Cooper Gillan >Assignee: Cooper Gillan >Priority: Major > Fix For: 2.0.0 > > > From [this PR > comment|https://github.com/apache/airflow/pull/7316#issuecomment-580946162]: > https://github.com/apache/airflow/blob/master/tests/providers/apache/hive/operators/test_hive.py#L29-L40 > This file contains tests for many unrelated components. I think it would be > useful to transfer some tests to other files. > Here is an example: PolideaInternal@a91b612 > To avoid conflicts you can do it in one PR, but it would be useful for each > move to be in a separate commit. This will facilitate the review. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[GitHub] [airflow] potiuk merged pull request #7469: [AIRFLOW-6842] Skip fixing ownership on Mac
potiuk merged pull request #7469: [AIRFLOW-6842] Skip fixing ownership on Mac URL: https://github.com/apache/airflow/pull/7469 This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[jira] [Commented] (AIRFLOW-6721) Reorganize and clean up Hive tests
[ https://issues.apache.org/jira/browse/AIRFLOW-6721?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17040418#comment-17040418 ] ASF GitHub Bot commented on AIRFLOW-6721: - mik-laj commented on pull request #7468: [AIRFLOW-6721] Organize hive tests URL: https://github.com/apache/airflow/pull/7468 This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org > Reorganize and clean up Hive tests > -- > > Key: AIRFLOW-6721 > URL: https://issues.apache.org/jira/browse/AIRFLOW-6721 > Project: Apache Airflow > Issue Type: Sub-task > Components: tests >Affects Versions: 1.10.7 >Reporter: Cooper Gillan >Assignee: Cooper Gillan >Priority: Major > Fix For: 2.0.0 > > > From [this PR > comment|https://github.com/apache/airflow/pull/7316#issuecomment-580946162]: > https://github.com/apache/airflow/blob/master/tests/providers/apache/hive/operators/test_hive.py#L29-L40 > This file contains tests for many unrelated components. I think it would be > useful to transfer some tests to other files. > Here is an example: PolideaInternal@a91b612 > To avoid conflicts you can do it in one PR, but it would be useful for each > move to be in a separate commit. This will facilitate the review. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Resolved] (AIRFLOW-6842) On Max checking for fix ownership returns space prefixed 0
[ https://issues.apache.org/jira/browse/AIRFLOW-6842?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jarek Potiuk resolved AIRFLOW-6842. --- Fix Version/s: 1.10.10 Resolution: Fixed > On Max checking for fix ownership returns space prefixed 0 > -- > > Key: AIRFLOW-6842 > URL: https://issues.apache.org/jira/browse/AIRFLOW-6842 > Project: Apache Airflow > Issue Type: Bug > Components: breeze >Affects Versions: 2.0.0 >Reporter: Jarek Potiuk >Priority: Major > Fix For: 1.10.10 > > > On Mac fixing ownership takes time and it is basically never needed. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Commented] (AIRFLOW-6842) On Max checking for fix ownership returns space prefixed 0
[ https://issues.apache.org/jira/browse/AIRFLOW-6842?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17040433#comment-17040433 ] ASF GitHub Bot commented on AIRFLOW-6842: - potiuk commented on pull request #7469: [AIRFLOW-6842] Skip fixing ownership on Mac URL: https://github.com/apache/airflow/pull/7469 This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org > On Max checking for fix ownership returns space prefixed 0 > -- > > Key: AIRFLOW-6842 > URL: https://issues.apache.org/jira/browse/AIRFLOW-6842 > Project: Apache Airflow > Issue Type: Bug > Components: breeze >Affects Versions: 2.0.0 >Reporter: Jarek Potiuk >Priority: Major > > On Mac fixing ownership takes time and it is basically never needed. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Commented] (AIRFLOW-6721) Reorganize and clean up Hive tests
[ https://issues.apache.org/jira/browse/AIRFLOW-6721?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17040428#comment-17040428 ] ASF subversion and git services commented on AIRFLOW-6721: -- Commit e91d615e02758da6b9f3cf6be502f0114d866ea3 in airflow's branch refs/heads/master from Cooper Gillan [ https://gitbox.apache.org/repos/asf?p=airflow.git;h=e91d615 ] [AIRFLOW-6721] Organize Apache Hive tests (#7468) * [AIRFLOW-6721] Move WebHdfsSensor tests to own module Move the WebHdfsSensor tests out of the operator tests and into a separate module. * [AIRFLOW-6721] Move TestHiveEnvironment to test module Movee the TestHiveEnvironment class and setUp method out to the hive test module so that it can be used in other hive tests. * [AIRFLOW-6721] Move Hive Stats tests to separate module Move skipped test for HiveStatsCollectionOperator to the test_hive_stats module to make it easier to locate. * [AIRFLOW-6721] Move NamedHivePartitionSensor tests Move the NamedHivePartitionSensor tests from the operators module into the test module for NamedHivePartitionSensor. * [AIRFLOW-6721] Move HivePartitionSensor tests Move the HivePartitionSensor tests from the operators module into the a new test module. * [AIRFLOW-6721] Move HiveToMySqlTransfer test Move the HiveToMySqlTransfer test in the operators module into the existing test module and use the shared test environment for these tests. * [AIRFLOW-6721] Move HdfsSensor test to separate module Move the HdfsSensor test in the operators module into a new test module and use the shared test environment. * [AIRFLOW-6721] Move Hive2SambaOperator test Move the Hive2SambaOperator test in the operators module into the existing test module and use the shared test environment. * [AIRFLOW-6721] Move Hive MetastorePartitionSensor test Move the MetastorePartitionSensor test from the operators module into a new module and use the shared test environment. * [AIRFLOW-6721] Move Hive PrestoToMySqlTransfer test Move the PrestoToMySqlTransfer test from the operators module into an existing test module and use shared Hive test environment. * [AIRFLOW-6721] Move PrestoCheckOperator test Move the PrestoCheckOperator test from the Hive operators module into a new Presto operators test module while still using the shared Hive test environment. * [AIRFLOW-6721] Move Hive SqlSensor test Move the SqlSensor in the Hive operators module into the existing SqlSensor test module and label it as such. Use the shared test environment for all tests, ensuring that original setUp properties are preserved for the existing tests. > Reorganize and clean up Hive tests > -- > > Key: AIRFLOW-6721 > URL: https://issues.apache.org/jira/browse/AIRFLOW-6721 > Project: Apache Airflow > Issue Type: Sub-task > Components: tests >Affects Versions: 1.10.7 >Reporter: Cooper Gillan >Assignee: Cooper Gillan >Priority: Major > Fix For: 2.0.0 > > > From [this PR > comment|https://github.com/apache/airflow/pull/7316#issuecomment-580946162]: > https://github.com/apache/airflow/blob/master/tests/providers/apache/hive/operators/test_hive.py#L29-L40 > This file contains tests for many unrelated components. I think it would be > useful to transfer some tests to other files. > Here is an example: PolideaInternal@a91b612 > To avoid conflicts you can do it in one PR, but it would be useful for each > move to be in a separate commit. This will facilitate the review. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Commented] (AIRFLOW-6721) Reorganize and clean up Hive tests
[ https://issues.apache.org/jira/browse/AIRFLOW-6721?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17040432#comment-17040432 ] ASF subversion and git services commented on AIRFLOW-6721: -- Commit e91d615e02758da6b9f3cf6be502f0114d866ea3 in airflow's branch refs/heads/master from Cooper Gillan [ https://gitbox.apache.org/repos/asf?p=airflow.git;h=e91d615 ] [AIRFLOW-6721] Organize Apache Hive tests (#7468) * [AIRFLOW-6721] Move WebHdfsSensor tests to own module Move the WebHdfsSensor tests out of the operator tests and into a separate module. * [AIRFLOW-6721] Move TestHiveEnvironment to test module Movee the TestHiveEnvironment class and setUp method out to the hive test module so that it can be used in other hive tests. * [AIRFLOW-6721] Move Hive Stats tests to separate module Move skipped test for HiveStatsCollectionOperator to the test_hive_stats module to make it easier to locate. * [AIRFLOW-6721] Move NamedHivePartitionSensor tests Move the NamedHivePartitionSensor tests from the operators module into the test module for NamedHivePartitionSensor. * [AIRFLOW-6721] Move HivePartitionSensor tests Move the HivePartitionSensor tests from the operators module into the a new test module. * [AIRFLOW-6721] Move HiveToMySqlTransfer test Move the HiveToMySqlTransfer test in the operators module into the existing test module and use the shared test environment for these tests. * [AIRFLOW-6721] Move HdfsSensor test to separate module Move the HdfsSensor test in the operators module into a new test module and use the shared test environment. * [AIRFLOW-6721] Move Hive2SambaOperator test Move the Hive2SambaOperator test in the operators module into the existing test module and use the shared test environment. * [AIRFLOW-6721] Move Hive MetastorePartitionSensor test Move the MetastorePartitionSensor test from the operators module into a new module and use the shared test environment. * [AIRFLOW-6721] Move Hive PrestoToMySqlTransfer test Move the PrestoToMySqlTransfer test from the operators module into an existing test module and use shared Hive test environment. * [AIRFLOW-6721] Move PrestoCheckOperator test Move the PrestoCheckOperator test from the Hive operators module into a new Presto operators test module while still using the shared Hive test environment. * [AIRFLOW-6721] Move Hive SqlSensor test Move the SqlSensor in the Hive operators module into the existing SqlSensor test module and label it as such. Use the shared test environment for all tests, ensuring that original setUp properties are preserved for the existing tests. > Reorganize and clean up Hive tests > -- > > Key: AIRFLOW-6721 > URL: https://issues.apache.org/jira/browse/AIRFLOW-6721 > Project: Apache Airflow > Issue Type: Sub-task > Components: tests >Affects Versions: 1.10.7 >Reporter: Cooper Gillan >Assignee: Cooper Gillan >Priority: Major > Fix For: 2.0.0 > > > From [this PR > comment|https://github.com/apache/airflow/pull/7316#issuecomment-580946162]: > https://github.com/apache/airflow/blob/master/tests/providers/apache/hive/operators/test_hive.py#L29-L40 > This file contains tests for many unrelated components. I think it would be > useful to transfer some tests to other files. > Here is an example: PolideaInternal@a91b612 > To avoid conflicts you can do it in one PR, but it would be useful for each > move to be in a separate commit. This will facilitate the review. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Commented] (AIRFLOW-6721) Reorganize and clean up Hive tests
[ https://issues.apache.org/jira/browse/AIRFLOW-6721?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17040427#comment-17040427 ] ASF subversion and git services commented on AIRFLOW-6721: -- Commit e91d615e02758da6b9f3cf6be502f0114d866ea3 in airflow's branch refs/heads/master from Cooper Gillan [ https://gitbox.apache.org/repos/asf?p=airflow.git;h=e91d615 ] [AIRFLOW-6721] Organize Apache Hive tests (#7468) * [AIRFLOW-6721] Move WebHdfsSensor tests to own module Move the WebHdfsSensor tests out of the operator tests and into a separate module. * [AIRFLOW-6721] Move TestHiveEnvironment to test module Movee the TestHiveEnvironment class and setUp method out to the hive test module so that it can be used in other hive tests. * [AIRFLOW-6721] Move Hive Stats tests to separate module Move skipped test for HiveStatsCollectionOperator to the test_hive_stats module to make it easier to locate. * [AIRFLOW-6721] Move NamedHivePartitionSensor tests Move the NamedHivePartitionSensor tests from the operators module into the test module for NamedHivePartitionSensor. * [AIRFLOW-6721] Move HivePartitionSensor tests Move the HivePartitionSensor tests from the operators module into the a new test module. * [AIRFLOW-6721] Move HiveToMySqlTransfer test Move the HiveToMySqlTransfer test in the operators module into the existing test module and use the shared test environment for these tests. * [AIRFLOW-6721] Move HdfsSensor test to separate module Move the HdfsSensor test in the operators module into a new test module and use the shared test environment. * [AIRFLOW-6721] Move Hive2SambaOperator test Move the Hive2SambaOperator test in the operators module into the existing test module and use the shared test environment. * [AIRFLOW-6721] Move Hive MetastorePartitionSensor test Move the MetastorePartitionSensor test from the operators module into a new module and use the shared test environment. * [AIRFLOW-6721] Move Hive PrestoToMySqlTransfer test Move the PrestoToMySqlTransfer test from the operators module into an existing test module and use shared Hive test environment. * [AIRFLOW-6721] Move PrestoCheckOperator test Move the PrestoCheckOperator test from the Hive operators module into a new Presto operators test module while still using the shared Hive test environment. * [AIRFLOW-6721] Move Hive SqlSensor test Move the SqlSensor in the Hive operators module into the existing SqlSensor test module and label it as such. Use the shared test environment for all tests, ensuring that original setUp properties are preserved for the existing tests. > Reorganize and clean up Hive tests > -- > > Key: AIRFLOW-6721 > URL: https://issues.apache.org/jira/browse/AIRFLOW-6721 > Project: Apache Airflow > Issue Type: Sub-task > Components: tests >Affects Versions: 1.10.7 >Reporter: Cooper Gillan >Assignee: Cooper Gillan >Priority: Major > Fix For: 2.0.0 > > > From [this PR > comment|https://github.com/apache/airflow/pull/7316#issuecomment-580946162]: > https://github.com/apache/airflow/blob/master/tests/providers/apache/hive/operators/test_hive.py#L29-L40 > This file contains tests for many unrelated components. I think it would be > useful to transfer some tests to other files. > Here is an example: PolideaInternal@a91b612 > To avoid conflicts you can do it in one PR, but it would be useful for each > move to be in a separate commit. This will facilitate the review. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[GitHub] [airflow] potiuk edited a comment on issue #7191: [AIRFLOW-4030] second attempt to add singularity to airflow
potiuk edited a comment on issue #7191: [AIRFLOW-4030] second attempt to add singularity to airflow URL: https://github.com/apache/airflow/pull/7191#issuecomment-588461323 > One quick note about the docs for pylint - the files are now changed to have pylint_main.sh and pylint_tests.sh. Also, this unexpectedly started that huge build process that failed before. Given a .pylintrc file in some root and running pylint, it would be great to see a much simpler (not wrapped / abstracted) run of pylint on a single file. I'd really recommend (again) pre-commits. They will run pylint on all modified files only. This is far better than having to run it manually - the will simply run pylint (or whatever of the other tests) automatically on all the files you modified when committing it. You don't even have to remember about it. And the reason for the build process is to make sure that you run it in the same environment as Travis CI and other people. The problems with tools like pylint is that the results depend often on installed libraries, pylint version etc. For example you need to have installed stubs for python libraries in order to do full static analysis. So there is huge variability depending on your local environment and you can simply get different results locally than on Travis. That's why we run pylint tests inside the docker container which mirrors this container that runs on Travis. This is the only way to make sure everyone runs in the same environment. It happened many times that people reported "I see different errors than those on Travis" - and it turned out that they have different pylint or python version. This is the real benefit of the "huge build". you start to appreciate it when your team of contributors is totally distributed, maintain their own environment and consists of 600 individuals - each with their own setup. Simply - it looks differently from the point of view of single person and differently from the point of view of team of few hundred people. Sometimes efficiency of team requires some sacrifice of efficiency of single individual. This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[GitHub] [airflow] potiuk commented on issue #7191: [AIRFLOW-4030] second attempt to add singularity to airflow
potiuk commented on issue #7191: [AIRFLOW-4030] second attempt to add singularity to airflow URL: https://github.com/apache/airflow/pull/7191#issuecomment-588461323 > One quick note about the docs for pylint - the files are now changed to have pylint_main.sh and pylint_tests.sh. Also, this unexpectedly started that huge build process that failed before. Given a .pylintrc file in some root and running pylint, it would be great to see a much simpler (not wrapped / abstracted) run of pylint on a single file. I'd really recommend (again) pre-commits. They will run pylint on all modified files only. This is far better than having to run it manually - the will simply run pylint (or whatever of the other tests) automatically on all the files you modified when committing it. You don't even have to remember about it. And the reason for the build process is to make sure that you run it in the same environment as Travis CI and other people. The problems with tools like pylint is that the results depend often on installed libraries, pylint version etc. For example you need to have installed stubs for python libraries in order to do full static analysis. So there is huge variability depending on your local environment and you can simply get different results locally than on Travis. That's why we run pylint tests inside the docker container which mirrors this container that runs on Travis. This is the only way to make sure everyone runs in the same environment. It happened many times that people reported "I see different errors than those on Travis" - and it turned out that they have different pylint or python version. This is the real benefit of the "huge build". you start to appreciate it when your team of contributors is totally distributed, maintain their own environment and consists of 600 individuals - each with their own setup. Simply - it looks differently from the point of view of single person and differently from the point of view of team of few hundred people. Sometimes efficiency of team require some sacrifice of efficiency of single individual. This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[GitHub] [airflow] mik-laj merged pull request #7468: [AIRFLOW-6721] Organize hive tests
mik-laj merged pull request #7468: [AIRFLOW-6721] Organize hive tests URL: https://github.com/apache/airflow/pull/7468 This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[GitHub] [airflow] mik-laj commented on issue #7468: [AIRFLOW-6721] Organize hive tests
mik-laj commented on issue #7468: [AIRFLOW-6721] Organize hive tests URL: https://github.com/apache/airflow/pull/7468#issuecomment-588458850 Can you create a ticket about `AIRFLOW_RUNALL_TESTS ` environment variable? This should be replaced by pytest markers. More information: https://github.com/apache/airflow/blob/master/TESTING.rst#airflow-integration-tests This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[jira] [Resolved] (AIRFLOW-6721) Reorganize and clean up Hive tests
[ https://issues.apache.org/jira/browse/AIRFLOW-6721?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Kamil Bregula resolved AIRFLOW-6721. Fix Version/s: 2.0.0 Resolution: Fixed > Reorganize and clean up Hive tests > -- > > Key: AIRFLOW-6721 > URL: https://issues.apache.org/jira/browse/AIRFLOW-6721 > Project: Apache Airflow > Issue Type: Sub-task > Components: tests >Affects Versions: 1.10.7 >Reporter: Cooper Gillan >Assignee: Cooper Gillan >Priority: Major > Fix For: 2.0.0 > > > From [this PR > comment|https://github.com/apache/airflow/pull/7316#issuecomment-580946162]: > https://github.com/apache/airflow/blob/master/tests/providers/apache/hive/operators/test_hive.py#L29-L40 > This file contains tests for many unrelated components. I think it would be > useful to transfer some tests to other files. > Here is an example: PolideaInternal@a91b612 > To avoid conflicts you can do it in one PR, but it would be useful for each > move to be in a separate commit. This will facilitate the review. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[GitHub] [airflow] vsoch commented on issue #7191: [AIRFLOW-4030] second attempt to add singularity to airflow
vsoch commented on issue #7191: [AIRFLOW-4030] second attempt to add singularity to airflow URL: https://github.com/apache/airflow/pull/7191#issuecomment-588448150 okay, I found plenty of examples and copied those, so I didn't need to run it manually. Let's hope that works! This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[GitHub] [airflow] vsoch commented on issue #7191: [AIRFLOW-4030] second attempt to add singularity to airflow
vsoch commented on issue #7191: [AIRFLOW-4030] second attempt to add singularity to airflow URL: https://github.com/apache/airflow/pull/7191#issuecomment-588442023 One quick note about the docs for pylint - the files are now changed to have pylint_main.sh and pylint_tests.sh. Also, this unexpectedly started that huge build process that failed before. Given a .pylintrc file in some root and running pylint, it would be great to see a much simpler (not wrapped / abstracted) run of pylint on a single file. This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[GitHub] [airflow] vsoch commented on issue #7191: [AIRFLOW-4030] second attempt to add singularity to airflow
vsoch commented on issue #7191: [AIRFLOW-4030] second attempt to add singularity to airflow URL: https://github.com/apache/airflow/pull/7191#issuecomment-588437467 okay, I'll try that! It does seem like I'm between a rock and a hard place, glad that you are okay with disabling pylint here. This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[GitHub] [airflow] potiuk commented on issue #7191: [AIRFLOW-4030] second attempt to add singularity to airflow
potiuk commented on issue #7191: [AIRFLOW-4030] second attempt to add singularity to airflow URL: https://github.com/apache/airflow/pull/7191#issuecomment-588435931 > How would you like me to address this? > > ``` > PATH=/root:/usr/local/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/opt/hive/bin:/usr/local/go/bin:/opt/gcloud/bin > > * Module airflow.providers.singularity.operators.singularity > > airflow/providers/singularity/operators/singularity.py:65:4: R0913: Too many arguments (12/10) (too-many-arguments) > ``` > > I had added working_dir to kwargs, but then during the test there was an error that the argument wasn't found. So I added it, but now I'm over. The pylint defats we have are 'reasonable' i.e work in most cases but if you have a justified case and do not think refactoring / using a structure instead of n-parameters you could disable pylint check - this specific check for this statement only by adding appropriate # pylint: disable comment. This is a sign from the programmer that this is a deliberate decision and theat it is conscious choice to disable tis rule. You can find plenty of examples in the code and more detailed explanation including some guidance on the decision making in the https://github.com/apache/airflow/blob/master/STATIC_CODE_CHECKS.rst in pylint section This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[jira] [Updated] (AIRFLOW-6846) Operator extra link function doesn't work without RBAC=True
[ https://issues.apache.org/jira/browse/AIRFLOW-6846?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Mark Zhang updated AIRFLOW-6846: Description: Operator extra link function does not work without enabling RBAC=True in airflow.cfg It'll be very nice if this useful feature can also function without enabling RBAC. And if there was a reason for this blocker, stating it in the documentations here: [https://airflow.apache.org/docs/stable/howto/define_extra_link.html] would be a good idea. was: Operator extra link function does not work without enabling `RBAC=True` in `airflow.cfg` It'll be very nice if this useful feature can also function without enabling `RBAC`. And if there was a reason for this blocker, stating it in the documentations here: [https://airflow.apache.org/docs/stable/howto/define_extra_link.html] would be a good idea. > Operator extra link function doesn't work without RBAC=True > --- > > Key: AIRFLOW-6846 > URL: https://issues.apache.org/jira/browse/AIRFLOW-6846 > Project: Apache Airflow > Issue Type: Bug > Components: documentation, operators, webserver >Affects Versions: 1.10.9 >Reporter: Mark Zhang >Priority: Minor > > Operator extra link function does not work without enabling RBAC=True in > airflow.cfg > It'll be very nice if this useful feature can also function without enabling > RBAC. And if there was a reason for this blocker, stating it in the > documentations here: > [https://airflow.apache.org/docs/stable/howto/define_extra_link.html] would > be a good idea. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Updated] (AIRFLOW-6846) Operator extra link function doesn't work without RBAC=True
[ https://issues.apache.org/jira/browse/AIRFLOW-6846?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Mark Zhang updated AIRFLOW-6846: Issue Type: Bug (was: Task) > Operator extra link function doesn't work without RBAC=True > --- > > Key: AIRFLOW-6846 > URL: https://issues.apache.org/jira/browse/AIRFLOW-6846 > Project: Apache Airflow > Issue Type: Bug > Components: documentation, operators, webserver >Affects Versions: 1.10.9 >Reporter: Mark Zhang >Priority: Minor > > Operator extra link function does not work without enabling `RBAC=True` in > `airflow.cfg` > > It'll be very nice if this useful feature can also function without enabling > `RBAC`. And if there was a reason for this blocker, stating it in the > documentations here: > [https://airflow.apache.org/docs/stable/howto/define_extra_link.html] would > be a good idea. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Updated] (AIRFLOW-6846) Operator extra link function doesn't work without RBAC=True
[ https://issues.apache.org/jira/browse/AIRFLOW-6846?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Mark Zhang updated AIRFLOW-6846: Description: Operator extra link function does not work without enabling `RBAC=True` in `airflow.cfg` It'll be very nice if this useful feature can also function without enabling `RBAC`. And if there was a reason for this blocker, stating it in the documentations here: [https://airflow.apache.org/docs/stable/howto/define_extra_link.html] would be a good idea. was: Operator extra link function does not work without enabling `RBAC=True` in `airflow.cfg` It'll be very nice if this useful feature can also function without enabling `RBAC`. And if there was a reason for this blocker, stating it in the documentations here: [https://airflow.apache.org/docs/stable/howto/define_extra_link.html] would be a good idea. > Operator extra link function doesn't work without RBAC=True > --- > > Key: AIRFLOW-6846 > URL: https://issues.apache.org/jira/browse/AIRFLOW-6846 > Project: Apache Airflow > Issue Type: Bug > Components: documentation, operators, webserver >Affects Versions: 1.10.9 >Reporter: Mark Zhang >Priority: Minor > > Operator extra link function does not work without enabling `RBAC=True` in > `airflow.cfg` > It'll be very nice if this useful feature can also function without enabling > `RBAC`. And if there was a reason for this blocker, stating it in the > documentations here: > [https://airflow.apache.org/docs/stable/howto/define_extra_link.html] would > be a good idea. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Created] (AIRFLOW-6846) Operator extra link function doesn't work without RBAC=True
Mark Zhang created AIRFLOW-6846: --- Summary: Operator extra link function doesn't work without RBAC=True Key: AIRFLOW-6846 URL: https://issues.apache.org/jira/browse/AIRFLOW-6846 Project: Apache Airflow Issue Type: Task Components: documentation, operators, webserver Affects Versions: 1.10.9 Reporter: Mark Zhang Operator extra link function does not work without enabling `RBAC=True` in `airflow.cfg` It'll be very nice if this useful feature can also function without enabling `RBAC`. And if there was a reason for this blocker, stating it in the documentations here: [https://airflow.apache.org/docs/stable/howto/define_extra_link.html] would be a good idea. -- This message was sent by Atlassian Jira (v8.3.4#803005)