Mike, I believe the /pcapGetter/getPcapsByIdentifiers endpoint exposes the fixed query option which we have covered. I agree with you that deprecating the metron-api module should be a goal of this feature.
On Wed, May 9, 2018 at 1:36 PM, Michael Miklavcic < michael.miklav...@gmail.com> wrote: > This looks like a pretty good start Ryan. Does the metadata endpoint cover > this https://github.com/apache/metron/tree/master/ > metron-platform/metron-api#the-pcapgettergetpcapsbyidentifiers-endpoint > from the original metron-api? If so, then we would be able to deprecate the > existing metron-api project. If we later go to micro-services, a pcap > module would spin back into the fold, but it would probably look different > from metron-api. > > I commented on the UI thread, but to reiterate for the purpose of backend > functionality here I don't believe there is a way to "PAUSE" or "SUSPEND" > jobs. That said, I think GET /api/v1/pcap/stop/<jobId> is sufficient for > the job management operations. > > On Wed, May 9, 2018 at 11:00 AM, Ryan Merriman <merrim...@gmail.com> > wrote: > > > Now that we are confident we can run submit a MR job from our current > REST > > application, is this the desired approach? Just want to confirm. > > > > Next I think we should map out what the REST interface will look like. > > Here are the endpoints I'm thinking about: > > > > GET /api/v1/pcap/metadata?basePath > > > > This endpoint will return metadata of pcap data stored in HDFS. This > would > > include pcap size, date ranges (how far back can I go), etc. It would > > accept an optional HDFS basePath parameter for cases where pcap data is > > stored in multiple places and/or different from the default location. > > > > POST /api/v1/pcap/query > > > > This endpoint would accept a pcap request, submit a pcap query job, and > > return a job id. The request would be an object containing the > parameters > > documented here: https://github.com/apache/metron/tree/master/ > > metron-platform/metron-pcap-backend#query-filter-utility. A query/job > > would be associated with a user that submits it. An exception will be > > returned for violating constraints like too many queries submitted, query > > parameters out of limits, etc. > > > > GET /api/v1/pcap/status/<jobId> > > > > This endpoint will return the status of a running job. I imagine this is > > just a proxy to the YARN REST api. We can discuss the implementation > > behind these endpoints later. > > > > GET /api/v1/pcap/stop/<jobId> > > > > This endpoint would kill a running pcap job. If the job has already > > completed this is a noop. > > > > GET /api/v1/pcap/list > > > > This endpoint will list a user's submitted pcap queries. Items in the > list > > would contain job id, status (is it finished?), start/end time, and > number > > of pages. Maybe there is some overlap with the status endpoint above and > > the status endpoint is not needed? > > > > GET /api/v1/pcap/pdml/<jobId>/<pageNumber> > > > > This endpoint will return pcap results for the given page in pdml format > ( > > https://wiki.wireshark.org/PDML). Are there other formats we want to > > support? > > > > GET /api/v1/pcap/raw/<jobId>/<pageNumber> > > > > This endpoint will allow a user to download raw pcap results for the > given > > page. > > > > DELETE /api/v1/pcap/<jobId> > > > > This endpoint will delete pcap query results. Not sure yet how this fits > > in with our broader cleanup strategy. > > > > This should get us started. What did I miss and what would you change > > about these? I did not include much detail related to security, cleanup > > strategy, or underlying implementation details but these are items we > > should discuss at some point. > > > > On Tue, May 8, 2018 at 5:38 PM, Michael Miklavcic < > > michael.miklav...@gmail.com> wrote: > > > > > Sweet! That's great news. The pom changes are a lot simpler than I > > > expected. Very nice. > > > > > > On Tue, May 8, 2018 at 4:35 PM, Ryan Merriman <merrim...@gmail.com> > > wrote: > > > > > > > Finally figured it out. Commit is here: > > > > https://github.com/merrimanr/incubator-metron/commit/ > > > > 22fe5e9ff3c167b42ebeb7a9f1000753a409aff1 > > > > > > > > It came down to figuring out the right combination of maven > > dependencies > > > > and passing in the HDP version to REST as a Java system property. I > > also > > > > included some HDFS setup tasks. I tested this in full dev and can > now > > > > successfully run a pcap query and get results. All you should have > to > > do > > > > is generate some pcap data first. > > > > > > > > On Tue, May 8, 2018 at 4:17 PM, Michael Miklavcic < > > > > michael.miklav...@gmail.com> wrote: > > > > > > > > > @Ryan - pulled your branch and experimented with a few things. In > > doing > > > > so, > > > > > it dawned on me that by adding the yarn and hadoop classpath, you > > > > probably > > > > > didn't introduce a new classpath issue, rather you probably just > > moved > > > > onto > > > > > the next classpath issue, ie hbase per your exception about hbase > > jaxb. > > > > > Anyhow, I put up a branch with some pom changes worth trying in > > > > conjunction > > > > > with invoking the rest app startup via "/usr/bin/yarn jar" > > > > > > > > > > https://github.com/mmiklavc/metron/tree/ryan-rest-test > > > > > > > > > > https://github.com/mmiklavc/metron/commit/ > > > 5ca23580fc6e043fafae2327c80b65 > > > > > b20ca1c0c9 > > > > > > > > > > Mike > > > > > > > > > > > > > > > On Tue, May 8, 2018 at 7:44 AM, Simon Elliston Ball < > > > > > si...@simonellistonball.com> wrote: > > > > > > > > > > > That would be a step closer to something more like a > micro-service > > > > > > architecture. However, I would want to make sure we think about > the > > > > > > operational complexity, and mpack implications of having another > > > server > > > > > > installed and running somewhere on the cluster (also, ssl, > > kerberos, > > > > etc > > > > > > etc requirements for that service). > > > > > > > > > > > > On 8 May 2018 at 14:27, Ryan Merriman <merrim...@gmail.com> > wrote: > > > > > > > > > > > > > +1 to having metron-api as it's own service and using a gateway > > > type > > > > > > > pattern. > > > > > > > > > > > > > > On Tue, May 8, 2018 at 8:13 AM, Otto Fowler < > > > ottobackwa...@gmail.com > > > > > > > > > > > > wrote: > > > > > > > > > > > > > > > Why not have metron-api as it’s own service and use a > ‘gateway’ > > > > type > > > > > > > > pattern in rest? > > > > > > > > > > > > > > > > > > > > > > > > On May 8, 2018 at 08:45:33, Ryan Merriman ( > merrim...@gmail.com > > ) > > > > > wrote: > > > > > > > > > > > > > > > > Moving the yarn classpath command earlier in the classpath > now > > > > gives > > > > > > this > > > > > > > > error: > > > > > > > > > > > > > > > > Caused by: java.lang.NoSuchMethodError: > > > > > > > > javax.servlet.ServletContext.getVirtualServerName()Ljava/ > > > > > lang/String; > > > > > > > > > > > > > > > > I will experiment with other combinations, I suspect we will > > need > > > > > > > > finer-grain control over the order. > > > > > > > > > > > > > > > > The grep matches class names inside jar files. I use this all > > the > > > > > time > > > > > > > and > > > > > > > > it's really useful. > > > > > > > > > > > > > > > > The metron-rest jar is already shaded. > > > > > > > > > > > > > > > > Reverse engineering the yarn jar command was the next thing I > > was > > > > > going > > > > > > > to > > > > > > > > try. Will let you know how it goes. > > > > > > > > > > > > > > > > On Tue, May 8, 2018 at 12:36 AM, Michael Miklavcic < > > > > > > > > michael.miklav...@gmail.com> wrote: > > > > > > > > > > > > > > > > > What order did you add the hadoop or yarn classpath? The > > > "shaded" > > > > > > > > package > > > > > > > > > stands out to me in this name "org.apache.hadoop.hbase.* > > > shaded* > > > > > > > > > .org.codehaus.jackson.jaxrs.JacksonJaxbJsonProvider." > Maybe > > > try > > > > > > adding > > > > > > > > > those packages earlier on the classpath. > > > > > > > > > > > > > > > > > > I think that find command needs a "jar tvf", otherwise > you're > > > > > looking > > > > > > > > for a > > > > > > > > > class name in jar file names. > > > > > > > > > > > > > > > > > > Have you tried shading the rest jar? > > > > > > > > > > > > > > > > > > I'd also look at the classpath you get when running "yarn > > jar" > > > to > > > > > > start > > > > > > > > the > > > > > > > > > existing pcap service, per the instructions in > > > > > metron-api/README.md. > > > > > > > > > > > > > > > > > > > > > > > > > > > On Mon, May 7, 2018 at 3:28 PM, Ryan Merriman < > > > > merrim...@gmail.com > > > > > > > > > > > > > > wrote: > > > > > > > > > > > > > > > > > > > To explore the idea of merging metron-api into > metron-rest > > > and > > > > > > > running > > > > > > > > > pcap > > > > > > > > > > queries inside our REST application, I created a simple > > test > > > > > here: > > > > > > > > > > https://github.com/merrimanr/incubator-metron/tree/pcap- > > > > > rest-test. > > > > > > A > > > > > > > > > > summary of what's included: > > > > > > > > > > > > > > > > > > > > - Added pcap as a dependency in the metron-rest pom.xml > > > > > > > > > > - Added a pcap query controller endpoint at > > > > > > > > > > http://node1:8082/swagger-ui. > html#!/pcap-query-controller/ > > > > > > > > > queryUsingGET > > > > > > > > > > - Added a pcap query service that runs a simple, > hardcoded > > > > query > > > > > > > > > > > > > > > > > > > > Generate some pcap data using pycapa ( > > > > > > > > > > https://github.com/apache/metron/tree/master/metron- > > > > > sensors/pycapa > > > > > > ) > > > > > > > > and > > > > > > > > > > the > > > > > > > > > > pcap topology ( > > > > > > > > > > https://github.com/apache/metron/tree/master/metron- > > > > > > > > > > platform/metron-pcap-backend#starting-the-topology). > > > > > > > > > > After this initial setup there should be data in HDFS at > > > > > > > > > > "/apps/metron/pcap". I believe this should be enough to > > > > exercise > > > > > > the > > > > > > > > > > issue. Just hit the endpoint referenced above. I tested > > this > > > in > > > > > an > > > > > > > > > > already running full dev by building and deploying the > > > > > metron-rest > > > > > > > > jar. > > > > > > > > > I > > > > > > > > > > did not rebuild full dev with this change but I would > still > > > > > expect > > > > > > it > > > > > > > > to > > > > > > > > > > work. Let me know if it doesn't. > > > > > > > > > > > > > > > > > > > > The first error I see when I hit this endpoint is: > > > > > > > > > > > > > > > > > > > > java.lang.NoClassDefFoundError: > > > > > > > > > > org/apache/hadoop/yarn/webapp/ > YarnJacksonJaxbJsonProvider. > > > > > > > > > > > > > > > > > > > > Here are the things I've tried so far: > > > > > > > > > > > > > > > > > > > > - Run the REST application with the YARN jar command > since > > > this > > > > > is > > > > > > > how > > > > > > > > > > all our other YARN/MR-related applications are started > > > > > (metron-api, > > > > > > > > > > MAAS, > > > > > > > > > > pcap query, etc). I wouldn't expect this to work since we > > > have > > > > > > > > > runtime > > > > > > > > > > dependencies on our shaded elasticsearch and parser jars > > and > > > > I'm > > > > > > not > > > > > > > > > > aware > > > > > > > > > > of a way to add additional jars to the classpath with the > > > YARN > > > > > jar > > > > > > > > > > command > > > > > > > > > > (is there a way?). Either way I get this error: > > > > > > > > > > > > > > > > > > > > 18/05/04 19:49:56 WARN reflections.Reflections: could not > > > > create > > > > > > Dir > > > > > > > > > using > > > > > > > > > > jarFile from url file:/usr/hdp/2.6.4.0-91/ > > > > hadoop/lib/ojdbc6.jar. > > > > > > > > > skipping. > > > > > > > > > > java.lang.NullPointerException > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > - I tried adding `yarn classpath` and `hadoop classpath` > to > > > the > > > > > > > > > > classpath in /usr/metron/0.4.3/bin/metron-rest.sh (REST > > > start > > > > > > > > > > script). I > > > > > > > > > > get this error: > > > > > > > > > > > > > > > > > > > > java.lang.ClassNotFoundException: > > > > > > > > > > org.apache.hadoop.hbase.shaded.org.codehaus.jackson. > > > > > > > > > > jaxrs.JacksonJaxbJsonProvider > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > - I searched for the class in the previous attempt but > > could > > > > not > > > > > > find > > > > > > > > > it > > > > > > > > > > in full dev: > > > > > > > > > > > > > > > > > > > > find / -name "*.jar" 2>/dev/null | xargs grep > > > > > > > > > > org/apache/hadoop/hbase/shaded/org/codehaus/jackson/ > > > > > > > > > > jaxrs/JacksonJaxbJsonProvider > > > > > > > > > > 2>/dev/null > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > - Further up in the stack trace I see the error happens > > when > > > > > > > > > initiating > > > > > > > > > > the org.apache.hadoop.yarn.util.timeline.TimelineUtils > > > class. > > > > I > > > > > > > > > tried > > > > > > > > > > setting "yarn.timeline-service.enabled" in Ambari to > false > > > and > > > > > > then > > > > > > > I > > > > > > > > > > get > > > > > > > > > > this error: > > > > > > > > > > > > > > > > > > > > Unable to parse > > > > > > > > > > '/hdp/apps/${hdp.version}/mapreduce/mapreduce.tar.gz#mr- > > > > > framework' > > > > > > > as > > > > > > > > a > > > > > > > > > > URI, check the setting for mapreduce.application. > > > > framework.path > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > - I've tried adding different hadoop, hbase, yarn and > > > mapreduce > > > > > > Maven > > > > > > > > > > dependencies without any success > > > > > > > > > > - hadoop-yarn-client > > > > > > > > > > - hadoop-yarn-common > > > > > > > > > > - hadoop-mapreduce-client-core > > > > > > > > > > - hadoop-yarn-server-common > > > > > > > > > > - hadoop-yarn-api > > > > > > > > > > - hbase-server > > > > > > > > > > > > > > > > > > > > I will keep exploring other possible solutions. Let me > know > > > if > > > > > > anyone > > > > > > > > > has > > > > > > > > > > any ideas. > > > > > > > > > > > > > > > > > > > > On Mon, May 7, 2018 at 9:02 AM, Otto Fowler < > > > > > > ottobackwa...@gmail.com > > > > > > > > > > > > > > > > > > wrote: > > > > > > > > > > > > > > > > > > > > > I can imagine a new generic service(s) capability whose > > > job ( > > > > > pun > > > > > > > > > > intended > > > > > > > > > > > ) is to > > > > > > > > > > > abstract the submittal, tracking, and storage of > results > > to > > > > > yarn. > > > > > > > > > > > > > > > > > > > > > > It would be extended with storage providers, queue > > > provider, > > > > > > > > possibly > > > > > > > > > > some > > > > > > > > > > > set of policies or rather strategies. > > > > > > > > > > > > > > > > > > > > > > The pcap ‘report’ would be a client to that service, > the > > > > > > > specializes > > > > > > > > > the > > > > > > > > > > > service operation for the way we want pcap to work. > > > > > > > > > > > > > > > > > > > > > > We can then re-use the generic service for other long > > > running > > > > > > yarn > > > > > > > > > > > things….. > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > On May 7, 2018 at 09:56:51, Otto Fowler ( > > > > > ottobackwa...@gmail.com > > > > > > ) > > > > > > > > > wrote: > > > > > > > > > > > > > > > > > > > > > > RE: Tracking v. users > > > > > > > > > > > > > > > > > > > > > > The submittal and tracking can associate the submitter > > with > > > > the > > > > > > > yarn > > > > > > > > > job > > > > > > > > > > > and track that, > > > > > > > > > > > regardless of the yarn credentials. > > > > > > > > > > > > > > > > > > > > > > IE> if all submittals and monitoring are by the same > yarn > > > > user > > > > > ( > > > > > > > > > Metron ) > > > > > > > > > > > from a single or > > > > > > > > > > > co-operative set of services, that service can maintain > > the > > > > > > > mapping. > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > On May 7, 2018 at 09:39:52, Ryan Merriman ( > > > > merrim...@gmail.com > > > > > ) > > > > > > > > wrote: > > > > > > > > > > > > > > > > > > > > > > Otto, your use case makes sense to me. We'll have to > > think > > > > > about > > > > > > > how > > > > > > > > to > > > > > > > > > > > manage the user to job relationships. I'm assuming YARN > > > jobs > > > > > will > > > > > > > be > > > > > > > > > > > submitted as the metron service user so YARN won't keep > > > track > > > > > of > > > > > > > > this > > > > > > > > > for > > > > > > > > > > > us. Is that assumption correct? Do you have any ideas > for > > > > doing > > > > > > > > that? > > > > > > > > > > > > > > > > > > > > > > Mike, I can start a feature branch and experiment with > > > > merging > > > > > > > > > metron-api > > > > > > > > > > > into metron-rest. That should allow us to collaborate > on > > > any > > > > > > issues > > > > > > > > or > > > > > > > > > > > challenges. Also, can you expand on your idea to manage > > > > > external > > > > > > > > > > > dependencies as a special module? That seems like a > very > > > > > > attractive > > > > > > > > > > option > > > > > > > > > > > to me. > > > > > > > > > > > > > > > > > > > > > > On Fri, May 4, 2018 at 8:39 AM, Otto Fowler < > > > > > > > ottobackwa...@gmail.com> > > > > > > > > > > > > > > > > > > > wrote: > > > > > > > > > > > > > > > > > > > > > > > From my response on the other thread, but applicable > to > > > the > > > > > > > > backend > > > > > > > > > > > stuff: > > > > > > > > > > > > > > > > > > > > > > > > "The PCAP Query seems more like PCAP Report to me. > You > > > are > > > > > > > > > generating a > > > > > > > > > > > > report based on parameters. > > > > > > > > > > > > That report is something that takes some time and > > > external > > > > > > > process > > > > > > > > to > > > > > > > > > > > > generate… ie you have to wait for it. > > > > > > > > > > > > > > > > > > > > > > > > I can almost imagine a flow where you: > > > > > > > > > > > > > > > > > > > > > > > > * Are in the AlertUI > > > > > > > > > > > > * Ask to generate a PCAP report based on some > selected > > > > > > > > > > alerts/meta-alert, > > > > > > > > > > > > possibly picking from on or more report ‘templates’ > > > > > > > > > > > > that have query options etc > > > > > > > > > > > > * The report request is ‘queued’, that is dispatched > to > > > be > > > > be > > > > > > > > > > > > executed/generated > > > > > > > > > > > > * You as a user have a ‘queue’ of your report > results, > > > and > > > > > when > > > > > > > > the > > > > > > > > > > > report > > > > > > > > > > > > is done it is queued there > > > > > > > > > > > > * We ‘monitor’ the report/queue press through the > yarn > > > > rest ( > > > > > > > > report > > > > > > > > > > > > info/meta has the yarn details ) > > > > > > > > > > > > * You can select the report from your queue and view > it > > > > > either > > > > > > in > > > > > > > > a > > > > > > > > > new > > > > > > > > > > > UI > > > > > > > > > > > > or custom component > > > > > > > > > > > > * You can then apply a different ‘view’ to the report > > or > > > > work > > > > > > > with > > > > > > > > > the > > > > > > > > > > > > report data > > > > > > > > > > > > * You can print / save etc > > > > > > > > > > > > * You can associate the report with the alerts ( > again > > in > > > > the > > > > > > > > report > > > > > > > > > > info > > > > > > > > > > > > ) with…. a ‘case’ or ‘ticket’ or investigation > > something > > > or > > > > > > other > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > We can introduce extensibility into the report > > templates, > > > > > > report > > > > > > > > > views > > > > > > > > > > ( > > > > > > > > > > > > thinks that work with the json data of the report ) > > > > > > > > > > > > > > > > > > > > > > > > Something like that.” > > > > > > > > > > > > > > > > > > > > > > > > Maybe we can do : > > > > > > > > > > > > > > > > > > > > > > > > template -> query parameters -> script => yarn info > > > > > > > > > > > > yarn info + query info + alert context + yarn status > => > > > > > report > > > > > > > > info > > > > > > > > > -> > > > > > > > > > > > > stored in a user’s ‘report queue’ > > > > > > > > > > > > report persistence added to report info > > > > > > > > > > > > metron-rest -> api to monitor the queue, read > results ( > > > > page > > > > > ), > > > > > > > > etc > > > > > > > > > etc > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > On May 4, 2018 at 09:23:39, Ryan Merriman ( > > > > > merrim...@gmail.com > > > > > > ) > > > > > > > > > wrote: > > > > > > > > > > > > > > > > > > > > > > > > I started a separate thread on Pcap UI considerations > > and > > > > > user > > > > > > > > > > > > requirements > > > > > > > > > > > > at Otto's request. This should help us keep these two > > > > related > > > > > > but > > > > > > > > > > > separate > > > > > > > > > > > > discussions focused. > > > > > > > > > > > > > > > > > > > > > > > > On Fri, May 4, 2018 at 7:19 AM, Michel Sumbul < > > > > > > > > > michelsum...@gmail.com> > > > > > > > > > > > > wrote: > > > > > > > > > > > > > > > > > > > > > > > > > Hello, > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > (Youhouuu my first reply on this kind of mail > > chain^^) > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > If I may, I would like to share my view on the > > > following > > > > 3 > > > > > > > > points. > > > > > > > > > > > > > > > > > > > > > > > > > > - Backend: > > > > > > > > > > > > > > > > > > > > > > > > > > The current metron-api is totally seperate, it will > > be > > > > > logic > > > > > > > for > > > > > > > > me > > > > > > > > > > to > > > > > > > > > > > > have > > > > > > > > > > > > > it at the same place as the others rest api. > > Especially > > > > > when > > > > > > > > more > > > > > > > > > > > > security > > > > > > > > > > > > > will be added, it will not be needed to do the job > > > twice. > > > > > > > > > > > > > The current implementation send back a pcap object > > > which > > > > > > still > > > > > > > > need > > > > > > > > > > to > > > > > > > > > > > > be > > > > > > > > > > > > > decoded. In the opensoc, the decoding was done with > > > > tshard > > > > > on > > > > > > > > the > > > > > > > > > > > > frontend. > > > > > > > > > > > > > It will be good to have this decoding happening > > > directly > > > > on > > > > > > the > > > > > > > > > > backend > > > > > > > > > > > > to > > > > > > > > > > > > > not create a load on frontend. An option will be to > > > > install > > > > > > > > tshark > > > > > > > > > on > > > > > > > > > > > > the > > > > > > > > > > > > > rest server and to use to convert the pcap to xml > and > > > > then > > > > > > to a > > > > > > > > > json > > > > > > > > > > > > that > > > > > > > > > > > > > will be send to the frontend. > > > > > > > > > > > > > > > > > > > > > > > > > > I tried to start directly the map/reduce job to > > search > > > > over > > > > > > all > > > > > > > > the > > > > > > > > > > > pcap > > > > > > > > > > > > > data from the rest server and as Ryan mention it, > we > > > had > > > > > > > > trouble. I > > > > > > > > > > > will > > > > > > > > > > > > > try to find back the error. > > > > > > > > > > > > > > > > > > > > > > > > > > Then in the POC, what we tried is to use the > > pcap_query > > > > > > script > > > > > > > > and > > > > > > > > > > this > > > > > > > > > > > > > work fine. I just modified it that he sends back > > > directly > > > > > the > > > > > > > > > job_id > > > > > > > > > > of > > > > > > > > > > > > > yarn and not waiting that the job is finished. Then > > it > > > > will > > > > > > > > allow > > > > > > > > > the > > > > > > > > > > > UI > > > > > > > > > > > > > and the rest server to know what the status of the > > > > research > > > > > > by > > > > > > > > > > querying > > > > > > > > > > > > the > > > > > > > > > > > > > yarn rest api. This will allow the UI and the rest > > > server > > > > > to > > > > > > be > > > > > > > > > async > > > > > > > > > > > > > without any blocking phase. What do you think about > > > that? > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > Having the job submitted directly from the code of > > the > > > > rest > > > > > > > > server > > > > > > > > > > will > > > > > > > > > > > > be > > > > > > > > > > > > > perfect, but it will need a lot of investigation I > > > think > > > > > (but > > > > > > > > I'm > > > > > > > > > not > > > > > > > > > > > > the > > > > > > > > > > > > > expert so I might be completely wrong ^^). > > > > > > > > > > > > > > > > > > > > > > > > > > We know that the pcap_query scritp work fine so why > > not > > > > > > calling > > > > > > > > it? > > > > > > > > > > Is > > > > > > > > > > > > it > > > > > > > > > > > > > that bad? (maybe stupid question, but I really > don’t > > > see > > > > a > > > > > > lot > > > > > > > > of > > > > > > > > > > > > drawback) > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > - Front end: > > > > > > > > > > > > > > > > > > > > > > > > > > Adding the the pcap search to the alert UI is, I > > think, > > > > the > > > > > > > > easiest > > > > > > > > > > way > > > > > > > > > > > > to > > > > > > > > > > > > > move forward. But indeed, it will then be the > “Alert > > UI > > > > and > > > > > > > > > > pcapquery”. > > > > > > > > > > > > > Maybe the name of the UI should just change to > > > something > > > > > like > > > > > > > > > > > > “Monitoring & > > > > > > > > > > > > > Investigation UI” ? > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > Is there any roadmap or plan for the different UI? > I > > > mean > > > > > did > > > > > > > > you > > > > > > > > > > > > already > > > > > > > > > > > > > had discussion on how you see the ui evolving with > > the > > > > new > > > > > > > > feature > > > > > > > > > > that > > > > > > > > > > > > > will come in the future? > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > - Microservices: > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > What do you mean exactly by microservices? Is it to > > > > > separate > > > > > > > all > > > > > > > > > the > > > > > > > > > > > > > features in different projects? Or something like > > > having > > > > > the > > > > > > > > > > different > > > > > > > > > > > > > components in container like kubernet? (again maybe > > > > stupid > > > > > > > > > question, > > > > > > > > > > > but > > > > > > > > > > > > I > > > > > > > > > > > > > don’t clearly understand what you mean J ) > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > Michel > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > -- > > > > > > -- > > > > > > simon elliston ball > > > > > > @sireb > > > > > > > > > > > > > > > > > > > > >