I'm on a role with questions.

I'm curious to see if I can relieve processing pressure by adding a new vm.

Would you know how I would go about it?

Also
I would like to pull data from sources instead of have the sources push data to 
my site. Have you come across this scenario?

F



Sent from my Bell Samsung device over Canada's largest network.


-------- Original message --------
From: Frank Horsfall <frankhorsf...@cunet.carleton.ca>
Date: 2017-09-06 10:51 PM (GMT-05:00)
To: user@metron.apache.org
Subject: Re: Clearing of data to start over

Also

Laurens you recommended to make 3 Kafka brokers but the install wizard would 
not let me.

As a result my node1 is the only broker currently.  Would this cause a 
bottleneck?

If so is there a method to install and configures the 2 additional brokers post 
initial install?

kindest regards

Frank



Sent from my Bell Samsung device over Canada's largest network.


-------- Original message --------
From: Frank Horsfall <frankhorsf...@cunet.carleton.ca>
Date: 2017-09-06 10:38 PM (GMT-05:00)
To: user@metron.apache.org
Subject: Re: Clearing of data to start over

Thanks Laurens and Nick.

I want to let the queues run over night to give us some possible insights into 
heap sizes etc.

I currently have 3 vms configured each with 8 cores  500 gigs of data capacity  
and 30 gigs of memory.

Elasticsearch has been configured with 10 gigs xmx.

I've set storm worker childopts at 7 gigs for now so it takes a while to max 
out and generate heap errors.

I deleted approx 6 million events and shut off the data generating apps.

The idea is to see how much will be processed overnight.

One thing that has me puzzled is why my bro app isn't emitting events. I double 
checked my config based on what's recommended but nothing is coming through. A 
mystery. lol


Also I kept some notes during the whole process and want to share them if you 
are interested.  let me know

Frank








Sent from my Bell Samsung device over Canada's largest network.


-------- Original message --------
From: Laurens Vets <laur...@daemon.be>
Date: 2017-09-06 6:17 PM (GMT-05:00)
To: user@metron.apache.org
Cc: Frank Horsfall <frankhorsf...@cunet.carleton.ca>
Subject: Re: Clearing of data to start over


Hi Frank,

If you all your queues (Kafka/Storm) are empty, the following should work:

- Deleting your elasticsearch indices: curl -X DELETE 
'http://localhost:9200/snort_index_*', curl -X DELETE 
'http://localhost:9200/yaf_index_*', etc...

- Deleting your Hadoop data:

Become the hdfs user: sudo su - hdfs
Show what's been indexed in Hadoop: hdfs dfs -ls /apps/metron/indexing/indexed/
Output should show the following probably:
/apps/metron/indexing/indexed/error
/apps/metron/indexing/indexed/snort
/apps/metron/indexing/indexed/yaf
...

You can remove these with:
hdfs dfs -rmr -skipTrash /apps/metron/indexing/indexed/error/
hdfs dfs -rmr -skipTrash /apps/metron/indexing/indexed/snort/

Or the individial files with

hdfs dfs -rmr -skipTrash /apps/metron/indexing/indexed/error/FILENAME


On 2017-09-06 13:59, Frank Horsfall wrote:
Hello all,
I have installed a 3 node system using the bare metal Centos 7 guideline.

https://cwiki.apache.org/confluence/display/METRON/Metron+0.4.0+with+HDP+2.5+bare-metal+install+on+Centos+7+with+MariaDB+for+Metron+REST

It has taken me a while to have all components working properly and I left the 
yaf,bro,snort apps running so quite a lot of data has been generated.  
Currently, I have almost 18 million events identified in Kibana. 16+ million 
are yaf based, and 2+ million are snort  …. 190 events are my new squid 
telemetry,  :).   It looks like it still has a while to go before it catches up 
to current day.   I recently shutdown the apps.


My questions are:


1.       Is there a way to wipe all my data and indices clean so that I may now 
begin with a fresh dataset?

2.       Is there a way to configure yaf so that its data is meaningful ? It is 
currently  creating what looks to be test data?

3.       I have commented out the test snort rule  but it is still generating 
the odd record which looks once again looks like test data. Can this be stopped 
as well?

Kindest regards,
Frank




Reply via email to