This looks like a pretty good start Ryan. Does the metadata endpoint cover
this https://github.com/apache/metron/tree/master/
metron-platform/metron-api#the-pcapgettergetpcapsbyidentifiers-endpoint
from the original metron-api? If so, then we would be able to deprecate the
existing metron-api project. If we later go to micro-services, a pcap
module would spin back into the fold, but it would probably look different
from metron-api.

I commented on the UI thread, but to reiterate for the purpose of backend
functionality here I don't believe there is a way to "PAUSE" or "SUSPEND"
jobs. That said, I think GET /api/v1/pcap/stop/<jobId> is sufficient for
the job management operations.

On Wed, May 9, 2018 at 11:00 AM, Ryan Merriman <merrim...@gmail.com> wrote:

> Now that we are confident we can run submit a MR job from our current REST
> application, is this the desired approach?  Just want to confirm.
>
> Next I think we should map out what the REST interface will look like.
> Here are the endpoints I'm thinking about:
>
> GET /api/v1/pcap/metadata?basePath
>
> This endpoint will return metadata of pcap data stored in HDFS.  This would
> include pcap size, date ranges (how far back can I go), etc.  It would
> accept an optional HDFS basePath parameter for cases where pcap data is
> stored in multiple places and/or different from the default location.
>
> POST /api/v1/pcap/query
>
> This endpoint would accept a pcap request, submit a pcap query job, and
> return a job id.  The request would be an object containing the parameters
> documented here:  https://github.com/apache/metron/tree/master/
> metron-platform/metron-pcap-backend#query-filter-utility.  A query/job
> would be associated with a user that submits it.  An exception will be
> returned for violating constraints like too many queries submitted, query
> parameters out of limits, etc.
>
> GET /api/v1/pcap/status/<jobId>
>
> This endpoint will return the status of a running job.  I imagine this is
> just a proxy to the YARN REST api.  We can discuss the implementation
> behind these endpoints later.
>
> GET /api/v1/pcap/stop/<jobId>
>
> This endpoint would kill a running pcap job.  If the job has already
> completed this is a noop.
>
> GET /api/v1/pcap/list
>
> This endpoint will list a user's submitted pcap queries.  Items in the list
> would contain job id, status (is it finished?), start/end time, and number
> of pages.  Maybe there is some overlap with the status endpoint above and
> the status endpoint is not needed?
>
> GET /api/v1/pcap/pdml/<jobId>/<pageNumber>
>
> This endpoint will return pcap results for the given page in pdml format (
> https://wiki.wireshark.org/PDML).  Are there other formats we want to
> support?
>
> GET /api/v1/pcap/raw/<jobId>/<pageNumber>
>
> This endpoint will allow a user to download raw pcap results for the given
> page.
>
> DELETE /api/v1/pcap/<jobId>
>
> This endpoint will delete pcap query results.  Not sure yet how this fits
> in with our broader cleanup strategy.
>
> This should get us started.  What did I miss and what would you change
> about these?  I did not include much detail related to security, cleanup
> strategy, or underlying implementation details but these are items we
> should discuss at some point.
>
> On Tue, May 8, 2018 at 5:38 PM, Michael Miklavcic <
> michael.miklav...@gmail.com> wrote:
>
> > Sweet! That's great news. The pom changes are a lot simpler than I
> > expected. Very nice.
> >
> > On Tue, May 8, 2018 at 4:35 PM, Ryan Merriman <merrim...@gmail.com>
> wrote:
> >
> > > Finally figured it out.  Commit is here:
> > > https://github.com/merrimanr/incubator-metron/commit/
> > > 22fe5e9ff3c167b42ebeb7a9f1000753a409aff1
> > >
> > > It came down to figuring out the right combination of maven
> dependencies
> > > and passing in the HDP version to REST as a Java system property.  I
> also
> > > included some HDFS setup tasks.  I tested this in full dev and can now
> > > successfully run a pcap query and get results.  All you should have to
> do
> > > is generate some pcap data first.
> > >
> > > On Tue, May 8, 2018 at 4:17 PM, Michael Miklavcic <
> > > michael.miklav...@gmail.com> wrote:
> > >
> > > > @Ryan - pulled your branch and experimented with a few things. In
> doing
> > > so,
> > > > it dawned on me that by adding the yarn and hadoop classpath, you
> > > probably
> > > > didn't introduce a new classpath issue, rather you probably just
> moved
> > > onto
> > > > the next classpath issue, ie hbase per your exception about hbase
> jaxb.
> > > > Anyhow, I put up a branch with some pom changes worth trying in
> > > conjunction
> > > > with invoking the rest app startup via "/usr/bin/yarn jar"
> > > >
> > > > https://github.com/mmiklavc/metron/tree/ryan-rest-test
> > > >
> > > > https://github.com/mmiklavc/metron/commit/
> > 5ca23580fc6e043fafae2327c80b65
> > > > b20ca1c0c9
> > > >
> > > > Mike
> > > >
> > > >
> > > > On Tue, May 8, 2018 at 7:44 AM, Simon Elliston Ball <
> > > > si...@simonellistonball.com> wrote:
> > > >
> > > > > That would be a step closer to something more like a micro-service
> > > > > architecture. However, I would want to make sure we think about the
> > > > > operational complexity, and mpack implications of having another
> > server
> > > > > installed and running somewhere on the cluster (also, ssl,
> kerberos,
> > > etc
> > > > > etc requirements for that service).
> > > > >
> > > > > On 8 May 2018 at 14:27, Ryan Merriman <merrim...@gmail.com> wrote:
> > > > >
> > > > > > +1 to having metron-api as it's own service and using a gateway
> > type
> > > > > > pattern.
> > > > > >
> > > > > > On Tue, May 8, 2018 at 8:13 AM, Otto Fowler <
> > ottobackwa...@gmail.com
> > > >
> > > > > > wrote:
> > > > > >
> > > > > > > Why not have metron-api as it’s own service and use a ‘gateway’
> > > type
> > > > > > > pattern in rest?
> > > > > > >
> > > > > > >
> > > > > > > On May 8, 2018 at 08:45:33, Ryan Merriman (merrim...@gmail.com
> )
> > > > wrote:
> > > > > > >
> > > > > > > Moving the yarn classpath command earlier in the classpath now
> > > gives
> > > > > this
> > > > > > > error:
> > > > > > >
> > > > > > > Caused by: java.lang.NoSuchMethodError:
> > > > > > > javax.servlet.ServletContext.getVirtualServerName()Ljava/
> > > > lang/String;
> > > > > > >
> > > > > > > I will experiment with other combinations, I suspect we will
> need
> > > > > > > finer-grain control over the order.
> > > > > > >
> > > > > > > The grep matches class names inside jar files. I use this all
> the
> > > > time
> > > > > > and
> > > > > > > it's really useful.
> > > > > > >
> > > > > > > The metron-rest jar is already shaded.
> > > > > > >
> > > > > > > Reverse engineering the yarn jar command was the next thing I
> was
> > > > going
> > > > > > to
> > > > > > > try. Will let you know how it goes.
> > > > > > >
> > > > > > > On Tue, May 8, 2018 at 12:36 AM, Michael Miklavcic <
> > > > > > > michael.miklav...@gmail.com> wrote:
> > > > > > >
> > > > > > > > What order did you add the hadoop or yarn classpath? The
> > "shaded"
> > > > > > > package
> > > > > > > > stands out to me in this name "org.apache.hadoop.hbase.*
> > shaded*
> > > > > > > > .org.codehaus.jackson.jaxrs.JacksonJaxbJsonProvider." Maybe
> > try
> > > > > adding
> > > > > > > > those packages earlier on the classpath.
> > > > > > > >
> > > > > > > > I think that find command needs a "jar tvf", otherwise you're
> > > > looking
> > > > > > > for a
> > > > > > > > class name in jar file names.
> > > > > > > >
> > > > > > > > Have you tried shading the rest jar?
> > > > > > > >
> > > > > > > > I'd also look at the classpath you get when running "yarn
> jar"
> > to
> > > > > start
> > > > > > > the
> > > > > > > > existing pcap service, per the instructions in
> > > > metron-api/README.md.
> > > > > > > >
> > > > > > > >
> > > > > > > > On Mon, May 7, 2018 at 3:28 PM, Ryan Merriman <
> > > merrim...@gmail.com
> > > > >
> > > > > > > wrote:
> > > > > > > >
> > > > > > > > > To explore the idea of merging metron-api into metron-rest
> > and
> > > > > > running
> > > > > > > > pcap
> > > > > > > > > queries inside our REST application, I created a simple
> test
> > > > here:
> > > > > > > > > https://github.com/merrimanr/incubator-metron/tree/pcap-
> > > > rest-test.
> > > > > A
> > > > > > > > > summary of what's included:
> > > > > > > > >
> > > > > > > > > - Added pcap as a dependency in the metron-rest pom.xml
> > > > > > > > > - Added a pcap query controller endpoint at
> > > > > > > > > http://node1:8082/swagger-ui.html#!/pcap-query-controller/
> > > > > > > > queryUsingGET
> > > > > > > > > - Added a pcap query service that runs a simple, hardcoded
> > > query
> > > > > > > > >
> > > > > > > > > Generate some pcap data using pycapa (
> > > > > > > > > https://github.com/apache/metron/tree/master/metron-
> > > > sensors/pycapa
> > > > > )
> > > > > > > and
> > > > > > > > > the
> > > > > > > > > pcap topology (
> > > > > > > > > https://github.com/apache/metron/tree/master/metron-
> > > > > > > > > platform/metron-pcap-backend#starting-the-topology).
> > > > > > > > > After this initial setup there should be data in HDFS at
> > > > > > > > > "/apps/metron/pcap". I believe this should be enough to
> > > exercise
> > > > > the
> > > > > > > > > issue. Just hit the endpoint referenced above. I tested
> this
> > in
> > > > an
> > > > > > > > > already running full dev by building and deploying the
> > > > metron-rest
> > > > > > > jar.
> > > > > > > > I
> > > > > > > > > did not rebuild full dev with this change but I would still
> > > > expect
> > > > > it
> > > > > > > to
> > > > > > > > > work. Let me know if it doesn't.
> > > > > > > > >
> > > > > > > > > The first error I see when I hit this endpoint is:
> > > > > > > > >
> > > > > > > > > java.lang.NoClassDefFoundError:
> > > > > > > > > org/apache/hadoop/yarn/webapp/YarnJacksonJaxbJsonProvider.
> > > > > > > > >
> > > > > > > > > Here are the things I've tried so far:
> > > > > > > > >
> > > > > > > > > - Run the REST application with the YARN jar command since
> > this
> > > > is
> > > > > > how
> > > > > > > > > all our other YARN/MR-related applications are started
> > > > (metron-api,
> > > > > > > > > MAAS,
> > > > > > > > > pcap query, etc). I wouldn't expect this to work since we
> > have
> > > > > > > > runtime
> > > > > > > > > dependencies on our shaded elasticsearch and parser jars
> and
> > > I'm
> > > > > not
> > > > > > > > > aware
> > > > > > > > > of a way to add additional jars to the classpath with the
> > YARN
> > > > jar
> > > > > > > > > command
> > > > > > > > > (is there a way?). Either way I get this error:
> > > > > > > > >
> > > > > > > > > 18/05/04 19:49:56 WARN reflections.Reflections: could not
> > > create
> > > > > Dir
> > > > > > > > using
> > > > > > > > > jarFile from url file:/usr/hdp/2.6.4.0-91/
> > > hadoop/lib/ojdbc6.jar.
> > > > > > > > skipping.
> > > > > > > > > java.lang.NullPointerException
> > > > > > > > >
> > > > > > > > >
> > > > > > > > > - I tried adding `yarn classpath` and `hadoop classpath` to
> > the
> > > > > > > > > classpath in /usr/metron/0.4.3/bin/metron-rest.sh (REST
> > start
> > > > > > > > > script). I
> > > > > > > > > get this error:
> > > > > > > > >
> > > > > > > > > java.lang.ClassNotFoundException:
> > > > > > > > > org.apache.hadoop.hbase.shaded.org.codehaus.jackson.
> > > > > > > > > jaxrs.JacksonJaxbJsonProvider
> > > > > > > > >
> > > > > > > > >
> > > > > > > > > - I searched for the class in the previous attempt but
> could
> > > not
> > > > > find
> > > > > > > > it
> > > > > > > > > in full dev:
> > > > > > > > >
> > > > > > > > > find / -name "*.jar" 2>/dev/null | xargs grep
> > > > > > > > > org/apache/hadoop/hbase/shaded/org/codehaus/jackson/
> > > > > > > > > jaxrs/JacksonJaxbJsonProvider
> > > > > > > > > 2>/dev/null
> > > > > > > > >
> > > > > > > > >
> > > > > > > > > - Further up in the stack trace I see the error happens
> when
> > > > > > > > initiating
> > > > > > > > > the org.apache.hadoop.yarn.util.timeline.TimelineUtils
> > class.
> > > I
> > > > > > > > tried
> > > > > > > > > setting "yarn.timeline-service.enabled" in Ambari to false
> > and
> > > > > then
> > > > > > I
> > > > > > > > > get
> > > > > > > > > this error:
> > > > > > > > >
> > > > > > > > > Unable to parse
> > > > > > > > > '/hdp/apps/${hdp.version}/mapreduce/mapreduce.tar.gz#mr-
> > > > framework'
> > > > > > as
> > > > > > > a
> > > > > > > > > URI, check the setting for mapreduce.application.
> > > framework.path
> > > > > > > > >
> > > > > > > > >
> > > > > > > > > - I've tried adding different hadoop, hbase, yarn and
> > mapreduce
> > > > > Maven
> > > > > > > > > dependencies without any success
> > > > > > > > > - hadoop-yarn-client
> > > > > > > > > - hadoop-yarn-common
> > > > > > > > > - hadoop-mapreduce-client-core
> > > > > > > > > - hadoop-yarn-server-common
> > > > > > > > > - hadoop-yarn-api
> > > > > > > > > - hbase-server
> > > > > > > > >
> > > > > > > > > I will keep exploring other possible solutions. Let me know
> > if
> > > > > anyone
> > > > > > > > has
> > > > > > > > > any ideas.
> > > > > > > > >
> > > > > > > > > On Mon, May 7, 2018 at 9:02 AM, Otto Fowler <
> > > > > ottobackwa...@gmail.com
> > > > > > >
> > > > > > > > > wrote:
> > > > > > > > >
> > > > > > > > > > I can imagine a new generic service(s) capability whose
> > job (
> > > > pun
> > > > > > > > > intended
> > > > > > > > > > ) is to
> > > > > > > > > > abstract the submittal, tracking, and storage of results
> to
> > > > yarn.
> > > > > > > > > >
> > > > > > > > > > It would be extended with storage providers, queue
> > provider,
> > > > > > > possibly
> > > > > > > > > some
> > > > > > > > > > set of policies or rather strategies.
> > > > > > > > > >
> > > > > > > > > > The pcap ‘report’ would be a client to that service, the
> > > > > > specializes
> > > > > > > > the
> > > > > > > > > > service operation for the way we want pcap to work.
> > > > > > > > > >
> > > > > > > > > > We can then re-use the generic service for other long
> > running
> > > > > yarn
> > > > > > > > > > things…..
> > > > > > > > > >
> > > > > > > > > >
> > > > > > > > > > On May 7, 2018 at 09:56:51, Otto Fowler (
> > > > ottobackwa...@gmail.com
> > > > > )
> > > > > > > > wrote:
> > > > > > > > > >
> > > > > > > > > > RE: Tracking v. users
> > > > > > > > > >
> > > > > > > > > > The submittal and tracking can associate the submitter
> with
> > > the
> > > > > > yarn
> > > > > > > > job
> > > > > > > > > > and track that,
> > > > > > > > > > regardless of the yarn credentials.
> > > > > > > > > >
> > > > > > > > > > IE> if all submittals and monitoring are by the same yarn
> > > user
> > > > (
> > > > > > > > Metron )
> > > > > > > > > > from a single or
> > > > > > > > > > co-operative set of services, that service can maintain
> the
> > > > > > mapping.
> > > > > > > > > >
> > > > > > > > > >
> > > > > > > > > >
> > > > > > > > > > On May 7, 2018 at 09:39:52, Ryan Merriman (
> > > merrim...@gmail.com
> > > > )
> > > > > > > wrote:
> > > > > > > > > >
> > > > > > > > > > Otto, your use case makes sense to me. We'll have to
> think
> > > > about
> > > > > > how
> > > > > > > to
> > > > > > > > > > manage the user to job relationships. I'm assuming YARN
> > jobs
> > > > will
> > > > > > be
> > > > > > > > > > submitted as the metron service user so YARN won't keep
> > track
> > > > of
> > > > > > > this
> > > > > > > > for
> > > > > > > > > > us. Is that assumption correct? Do you have any ideas for
> > > doing
> > > > > > > that?
> > > > > > > > > >
> > > > > > > > > > Mike, I can start a feature branch and experiment with
> > > merging
> > > > > > > > metron-api
> > > > > > > > > > into metron-rest. That should allow us to collaborate on
> > any
> > > > > issues
> > > > > > > or
> > > > > > > > > > challenges. Also, can you expand on your idea to manage
> > > > external
> > > > > > > > > > dependencies as a special module? That seems like a very
> > > > > attractive
> > > > > > > > > option
> > > > > > > > > > to me.
> > > > > > > > > >
> > > > > > > > > > On Fri, May 4, 2018 at 8:39 AM, Otto Fowler <
> > > > > > ottobackwa...@gmail.com>
> > > > > > >
> > > > > > > > > > wrote:
> > > > > > > > > >
> > > > > > > > > > > From my response on the other thread, but applicable to
> > the
> > > > > > > backend
> > > > > > > > > > stuff:
> > > > > > > > > > >
> > > > > > > > > > > "The PCAP Query seems more like PCAP Report to me. You
> > are
> > > > > > > > generating a
> > > > > > > > > > > report based on parameters.
> > > > > > > > > > > That report is something that takes some time and
> > external
> > > > > > process
> > > > > > > to
> > > > > > > > > > > generate… ie you have to wait for it.
> > > > > > > > > > >
> > > > > > > > > > > I can almost imagine a flow where you:
> > > > > > > > > > >
> > > > > > > > > > > * Are in the AlertUI
> > > > > > > > > > > * Ask to generate a PCAP report based on some selected
> > > > > > > > > alerts/meta-alert,
> > > > > > > > > > > possibly picking from on or more report ‘templates’
> > > > > > > > > > > that have query options etc
> > > > > > > > > > > * The report request is ‘queued’, that is dispatched to
> > be
> > > be
> > > > > > > > > > > executed/generated
> > > > > > > > > > > * You as a user have a ‘queue’ of your report results,
> > and
> > > > when
> > > > > > > the
> > > > > > > > > > report
> > > > > > > > > > > is done it is queued there
> > > > > > > > > > > * We ‘monitor’ the report/queue press through the yarn
> > > rest (
> > > > > > > report
> > > > > > > > > > > info/meta has the yarn details )
> > > > > > > > > > > * You can select the report from your queue and view it
> > > > either
> > > > > in
> > > > > > > a
> > > > > > > > new
> > > > > > > > > > UI
> > > > > > > > > > > or custom component
> > > > > > > > > > > * You can then apply a different ‘view’ to the report
> or
> > > work
> > > > > > with
> > > > > > > > the
> > > > > > > > > > > report data
> > > > > > > > > > > * You can print / save etc
> > > > > > > > > > > * You can associate the report with the alerts ( again
> in
> > > the
> > > > > > > report
> > > > > > > > > info
> > > > > > > > > > > ) with…. a ‘case’ or ‘ticket’ or investigation
> something
> > or
> > > > > other
> > > > > > > > > > >
> > > > > > > > > > >
> > > > > > > > > > > We can introduce extensibility into the report
> templates,
> > > > > report
> > > > > > > > views
> > > > > > > > > (
> > > > > > > > > > > thinks that work with the json data of the report )
> > > > > > > > > > >
> > > > > > > > > > > Something like that.”
> > > > > > > > > > >
> > > > > > > > > > > Maybe we can do :
> > > > > > > > > > >
> > > > > > > > > > > template -> query parameters -> script => yarn info
> > > > > > > > > > > yarn info + query info + alert context + yarn status =>
> > > > report
> > > > > > > info
> > > > > > > > ->
> > > > > > > > > > > stored in a user’s ‘report queue’
> > > > > > > > > > > report persistence added to report info
> > > > > > > > > > > metron-rest -> api to monitor the queue, read results (
> > > page
> > > > ),
> > > > > > > etc
> > > > > > > > etc
> > > > > > > > > > >
> > > > > > > > > > >
> > > > > > > > > > > On May 4, 2018 at 09:23:39, Ryan Merriman (
> > > > merrim...@gmail.com
> > > > > )
> > > > > > > > wrote:
> > > > > > > > > > >
> > > > > > > > > > > I started a separate thread on Pcap UI considerations
> and
> > > > user
> > > > > > > > > > > requirements
> > > > > > > > > > > at Otto's request. This should help us keep these two
> > > related
> > > > > but
> > > > > > > > > > separate
> > > > > > > > > > > discussions focused.
> > > > > > > > > > >
> > > > > > > > > > > On Fri, May 4, 2018 at 7:19 AM, Michel Sumbul <
> > > > > > > > michelsum...@gmail.com>
> > > > > > > > > > > wrote:
> > > > > > > > > > >
> > > > > > > > > > > > Hello,
> > > > > > > > > > > >
> > > > > > > > > > > >
> > > > > > > > > > > >
> > > > > > > > > > > > (Youhouuu my first reply on this kind of mail
> chain^^)
> > > > > > > > > > > >
> > > > > > > > > > > >
> > > > > > > > > > > >
> > > > > > > > > > > > If I may, I would like to share my view on the
> > following
> > > 3
> > > > > > > points.
> > > > > > > > > > > >
> > > > > > > > > > > > - Backend:
> > > > > > > > > > > >
> > > > > > > > > > > > The current metron-api is totally seperate, it will
> be
> > > > logic
> > > > > > for
> > > > > > > me
> > > > > > > > > to
> > > > > > > > > > > have
> > > > > > > > > > > > it at the same place as the others rest api.
> Especially
> > > > when
> > > > > > > more
> > > > > > > > > > > security
> > > > > > > > > > > > will be added, it will not be needed to do the job
> > twice.
> > > > > > > > > > > > The current implementation send back a pcap object
> > which
> > > > > still
> > > > > > > need
> > > > > > > > > to
> > > > > > > > > > > be
> > > > > > > > > > > > decoded. In the opensoc, the decoding was done with
> > > tshard
> > > > on
> > > > > > > the
> > > > > > > > > > > frontend.
> > > > > > > > > > > > It will be good to have this decoding happening
> > directly
> > > on
> > > > > the
> > > > > > > > > backend
> > > > > > > > > > > to
> > > > > > > > > > > > not create a load on frontend. An option will be to
> > > install
> > > > > > > tshark
> > > > > > > > on
> > > > > > > > > > > the
> > > > > > > > > > > > rest server and to use to convert the pcap to xml and
> > > then
> > > > > to a
> > > > > > > > json
> > > > > > > > > > > that
> > > > > > > > > > > > will be send to the frontend.
> > > > > > > > > > > >
> > > > > > > > > > > > I tried to start directly the map/reduce job to
> search
> > > over
> > > > > all
> > > > > > > the
> > > > > > > > > > pcap
> > > > > > > > > > > > data from the rest server and as Ryan mention it, we
> > had
> > > > > > > trouble. I
> > > > > > > > > > will
> > > > > > > > > > > > try to find back the error.
> > > > > > > > > > > >
> > > > > > > > > > > > Then in the POC, what we tried is to use the
> pcap_query
> > > > > script
> > > > > > > and
> > > > > > > > > this
> > > > > > > > > > > > work fine. I just modified it that he sends back
> > directly
> > > > the
> > > > > > > > job_id
> > > > > > > > > of
> > > > > > > > > > > > yarn and not waiting that the job is finished. Then
> it
> > > will
> > > > > > > allow
> > > > > > > > the
> > > > > > > > > > UI
> > > > > > > > > > > > and the rest server to know what the status of the
> > > research
> > > > > by
> > > > > > > > > querying
> > > > > > > > > > > the
> > > > > > > > > > > > yarn rest api. This will allow the UI and the rest
> > server
> > > > to
> > > > > be
> > > > > > > > async
> > > > > > > > > > > > without any blocking phase. What do you think about
> > that?
> > > > > > > > > > > >
> > > > > > > > > > > >
> > > > > > > > > > > >
> > > > > > > > > > > > Having the job submitted directly from the code of
> the
> > > rest
> > > > > > > server
> > > > > > > > > will
> > > > > > > > > > > be
> > > > > > > > > > > > perfect, but it will need a lot of investigation I
> > think
> > > > (but
> > > > > > > I'm
> > > > > > > > not
> > > > > > > > > > > the
> > > > > > > > > > > > expert so I might be completely wrong ^^).
> > > > > > > > > > > >
> > > > > > > > > > > > We know that the pcap_query scritp work fine so why
> not
> > > > > calling
> > > > > > > it?
> > > > > > > > > Is
> > > > > > > > > > > it
> > > > > > > > > > > > that bad? (maybe stupid question, but I really don’t
> > see
> > > a
> > > > > lot
> > > > > > > of
> > > > > > > > > > > drawback)
> > > > > > > > > > > >
> > > > > > > > > > > >
> > > > > > > > > > > >
> > > > > > > > > > > > - Front end:
> > > > > > > > > > > >
> > > > > > > > > > > > Adding the the pcap search to the alert UI is, I
> think,
> > > the
> > > > > > > easiest
> > > > > > > > > way
> > > > > > > > > > > to
> > > > > > > > > > > > move forward. But indeed, it will then be the “Alert
> UI
> > > and
> > > > > > > > > pcapquery”.
> > > > > > > > > > > > Maybe the name of the UI should just change to
> > something
> > > > like
> > > > > > > > > > > “Monitoring &
> > > > > > > > > > > > Investigation UI” ?
> > > > > > > > > > > >
> > > > > > > > > > > >
> > > > > > > > > > > >
> > > > > > > > > > > > Is there any roadmap or plan for the different UI? I
> > mean
> > > > did
> > > > > > > you
> > > > > > > > > > > already
> > > > > > > > > > > > had discussion on how you see the ui evolving with
> the
> > > new
> > > > > > > feature
> > > > > > > > > that
> > > > > > > > > > > > will come in the future?
> > > > > > > > > > > >
> > > > > > > > > > > >
> > > > > > > > > > > >
> > > > > > > > > > > > - Microservices:
> > > > > > > > > > > >
> > > > > > > > > > > >
> > > > > > > > > > > >
> > > > > > > > > > > > What do you mean exactly by microservices? Is it to
> > > > separate
> > > > > > all
> > > > > > > > the
> > > > > > > > > > > > features in different projects? Or something like
> > having
> > > > the
> > > > > > > > > different
> > > > > > > > > > > > components in container like kubernet? (again maybe
> > > stupid
> > > > > > > > question,
> > > > > > > > > > but
> > > > > > > > > > > I
> > > > > > > > > > > > don’t clearly understand what you mean J )
> > > > > > > > > > > >
> > > > > > > > > > > >
> > > > > > > > > > > >
> > > > > > > > > > > > Michel
> > > > > > > > > > > >
> > > > > > > > > > >
> > > > > > > > > > >
> > > > > > > > > >
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > > >
> > > > >
> > > > > --
> > > > > --
> > > > > simon elliston ball
> > > > > @sireb
> > > > >
> > > >
> > >
> >
>

Reply via email to