Hi, all Thanks for your reply and your helpful context.
I think we can classify the test of the connectors(adapters and sinks) into the following levels: * Unit tests: Testing only for individual modules or classes. For unit tests of connectors, there is no need to start the third-party container service. Examples are configuration tests(maybe connector client part only), helper class tests,s or some small validation tests, etc. We can use mock if needed. * Integration tests: Testing for those who need to start third-party services. I think this should be the main part of testing connectors. Because we can easily test a lot of things here. Examples are security(authentication) tests, configuration(both for the client part and server part) tests, data format validation tests, fuzz testing(we can simulate a greater diversity of data through the operation of third-party components), or any other component-specific tests. * E2E test: In this test, we can simulate the user's behavior more completely. We can use pre-defined data to test the basic connector functionality. In this concept, most of the functions of connectors can be covered in the integration tests. > I'm a little more in favor of taking the low-level API approach as it gives > us fewer dependencies Just to make sure, the fewer dependencies you mentioned here are from the streampipes like the streampipes backend service or other dependencies, right? In this case, I’m +1 for this idea. > For the PR you used the method ‘getNElements’. Which makes completely sense, > because it is a synchronous method call, however this method is only > available for generic adapters. Thanks for your helpful information. In this PR, I haven't found a better way to get the data yet, so I used the schema guessing method to get the data. > I see two options to get the resulting events from the adapter: Great! I am tring to investigate these options. > We can also think about refactoring the connect API. > I'm willing to help in improving the REST communication for tests and I will > also take a look in improving the StreamPipes Connect interfaces to ease the > setup of integration tests. +1 and thank you all. I’m also willing to help with that. > Do you want your PR merged into dev now or should we create a branch where we > work on the integration tests? I think we can merge that PR if you think there is no problem. We can then build on this foundation for subsequent development. Thanks, Zike Yang On Sun, Oct 16, 2022 at 8:21 PM Dominik Riemer <[email protected]> wrote: > > Hi Zike, > thanks a lot! This is very important work. > I'm willing to help in improving the REST communication for tests and I will > also take a look in improving the StreamPipes Connect interfaces to ease the > setup of integration tests. > > Do you want your PR merged into dev now or should we create a branch where we > work on the integration tests? > > Cheers > Dominik > > On 2022/10/12 16:17:38 Zike Yang wrote: > > Hi all, > > > > Thanks for your replies. I am very glad that all of you are involved > > in this discussion. > > > > I will try to answer your replies here: > > > > > Do you have an idea how we can realize all the user input in the > > > integration tests? > > Regarding the input data for tests, I haven't very clear ideas. This > > is an interesting point to investigate. > > > > The integration test will interact directly with the > > streampipes-backend. Although we can’t test these components at the UI > > level in the integration test, interacting via the backend using REST > > protocol should be able to simulate all user input. > > > > > Maybe it is even possible to re-use those or similar configuration files > > Agree. This will probably be a point of improvement. > > > > > How can we deal with third party dependencies? > > We could implement a TestContainer class for each third-party > > component. And we manage all dependencies for that component in that > > class. We are using the same approach for testing the pulsar sinks and > > pulsar sources component in the Apache Pulsar project.[0] > > > > > Maybe we can start with a rather simple example and based on that > > > together find a way to implement integration testing for StreamPipes? > > Great! > > > > > what do you think would be a good first step? > > I have created a PR[1] to demonstrate the integration test framework. > > And I have added an example for testing the Pulsar adapter which I am > > most familiar with. > > > > The PulsarContainer class manages the pulsar service. And we would > > implement the Tester for each component. PulsarAdapterTester is the > > tester for the pulsar adapter. We define some testing steps in the > > tester. And we execute the actual test in the AdaptersTest class by > > calling the third-party testers. > > > > This is a very simple and basic PR. I haven’t dealt with the part of > > interaction with the streampipes-backend. Actually, we need to create, > > interact, and delete the adapter through the backend instead of using > > the specific adapter class. The test in this PR needs to be optimized > > in the future. > > > > Regarding talking with the backend, I want to find the easiest way to > > create the adapter through the REST protocol. I tried to do that using > > this API[2]. But I found that the AdapterDescription is too complex > > for me. it would be appreciated If someone could help deal with that > > or guide me. > > > > I would also like to discuss a simpler way to take the loaded data > > from the adapter for testing through the REST. I haven't found a > > suitable approach or interface. > > > > And thank you all again. Please feel free to share your thoughts. > > > > [0] > > https://github.com/apache/pulsar/tree/master/tests/integration/src/test/java/org/apache/pulsar/tests/integration/io > > [1] https://github.com/apache/incubator-streampipes/pull/117 > > [2] > > https://github.com/apache/incubator-streampipes/blob/dev/streampipes-rest/src/main/java/org/apache/streampipes/rest/impl/connect/AdapterResource.java#L51 > > > > Best, > > Zike Yang > > > > On Wed, Oct 12, 2022 at 2:47 PM Dominik Riemer > > <[email protected]> wrote: > > > > > > Hi, > > > sounds great! > > > That's a very important issue and a module for integration tests makes > > > totally sense - would be cool to work towards a first example and then > > > extend this to all adapters, processors and sinks. > > > > > > @Zike Yang what do you think would be a good first step? > > > > > > Cheers > > > Dominik > > > > > > -----Original Message----- > > > From: Tim <[email protected]> > > > Sent: Tuesday, October 11, 2022 9:58 PM > > > To: [email protected] > > > Subject: Re: AW: [DISCUSS] Integration test for adapters and sinks > > > > > > Hi, > > > > > > Thank you for initiating this discussion and already providing a solid > > > foundation for integration testing for StreamPipes. > > > I completely agree with you and Philipp that testing (both in terms of > > > unit tests and integration tests) is a weak point in our codebase and has > > > been treated stepmotherly so far. > > > It would be great to have integration tests for all our adapters and > > > sinks. > > > Maybe we can start with a rather simple example and based on that > > > together find a way to implement integration testing for StreamPipes? > > > > > > I would be happy to contribute to this effort to some extent. > > > > > > Best > > > Tim > > > > > > Am 11.10.2022 19:15 schrieb Philipp Zehnder: > > > > Hi Zike Yang, > > > > > > > > thanks a lot for opening the discussion. I really like this idea! > > > > We started with the e2e tests to have some basic tests, but we > > > > definitely need way more test coverage and any help is highly > > > > appreciated. > > > > To have some kind of integration or unit tests for adapters and > > > > processors would be awesome, especially because users can provide a > > > > wide range of input configurations. > > > > > > > > For the e2e tests we tried to provide as much as possible via > > > > configuration files see [1]. They describe the configuration for the > > > > processors, the input data and the expected output data. This is > > > > required to also test more complex patterns in the event stream. Do > > > > you have an idea how we can realize all the user input in the > > > > integration tests? > > > > Maybe it is even possible to re-use those or similar configuration > > > > files for both the integration and the e2e tests of processors. I > > > > expect the integration tests to be much faster then the e2e tests, but > > > > for the e2e tests we can ensure that the user can use the GUI for all > > > > the input. > > > > > > > > How can we deal with third party dependencies? Especially for adapters > > > > and data sinks we require other services that must be configured > > > > accordingly. Do you have any experience with that how to do it best? > > > > > > > > I am really looking forward to discuss this further with you. > > > > > > > > Cheers, > > > > Philipp > > > > > > > > [1] > > > > https://github.com/apache/incubator-streampipes/tree/dev/ui/cypress/fi > > > > xtures/pipelineElement > > > > > > > > Von: Zike Yang <[email protected]> > > > > Datum: Dienstag, 11. Oktober 2022 um 17:31 > > > > An: [email protected] <[email protected]> > > > > Betreff: [DISCUSS] Integration test for adapters and sinks Hi, all > > > > > > > > I want to use this thread to discuss the integration tests for > > > > adapters and sinks. > > > > > > > > Currently, there seems to be very little testing for adapters and > > > > sinks. It’s not convenient to develop and fix them. We already have > > > > the e2e tests for these third-party components[0]. But this requires > > > > starting all third-party docker containers before running e2e tests. > > > > If there are many adapters and sinks components(actually we already > > > > have), it will lead to inconvenient testing. It also seems > > > > inconvenient to develop and debug them. I wonder if there is a better > > > > way to optimize this testing approach for third-party components. > > > > > > > > I would like to propose adding the integration test for adapters and > > > > sinks. And we could still use the e2e tests to do some small smoke > > > > tests. > > > > > > > > We could add a new module called streampipes-integration-tests. In > > > > this module, we do all integration tests for all adapters and sinks > > > > and perhaps some other components. For each test, we could start the > > > > docker container of the third-party component using Testcontainers[1]. > > > > And we use the streampipes-backend to create, interact, and delete the > > > > component through the Rest protocol. We could abstract some common > > > > tests and utilities for all adapters and sinks. > > > > > > > > Using the streampipes-integration-tests, we can test a large number of > > > > adapters and sinks and be able to clean up these docker container > > > > resources in a timely manner. It also facilitates the development and > > > > debugging. It is still possible to test these components at a higher > > > > level(at the backend level). > > > > > > > > In addition, regarding the unit test, we could also add some unit > > > > tests for each adapter and sink at their corresponding modules if > > > > necessary. > > > > > > > > Please feel free to share your thoughts. And I'm interested in making > > > > it happen. > > > > > > > > [0] > > > > https://github.com/apache/incubator-streampipes/tree/dev/ui/cypress/te > > > > sts/thirdparty > > > > [1] https://www.testcontainers.org/ > > > > > > > > Thanks, > > > > Zike Yang > >
