This is not complete in any way... updating slowly... its painful
First thing we need to do is setup the storage area for all of our pcap or json files. Later we will use filebeat and connect to the file share storage and ship those files to logstash for processing into elasticsearch.
Follow this how-to from Microsoft:
https://docs.microsoft.com/en-us/azure/storage/files/storage-how-to-create-file-share
Install filebeat:
https://www.elastic.co/guide/en/beats/filebeat/current/filebeat-installation.html
Configure filebeat.yml with this following information:
Filebeat.yml
filebeat:
prospectors:
- type: log
paths:
- /mnt/log/*.json
processors:
- drop_fields:
when:
has_fields: ['host']
fields:
- 'host'
output:
logstash:
hosts:
- IPADDRESS TO SELKS:5044
Mount Azure fileshare to or your own path
/mnt/log
Start filebeat
service filebeat start
Validate filebeat is connected to SELKS logstash
tail -f /var/log/filebeat/filebea
SELKS probe for Azure (depends on level of detail, more to come)
Options: -daemonlogger -rpcap -Suricata
Add ansible scripts to reboot/update probes.
Quick and dirty, more to come:
In a nutshell: Install a lightweight server (CentOS or Ubuntu), this is required because of the SMB3 requirements and if you want more security, then SMB3>SMB2+. Currently SELKS is built on Debian 9 but the kernel at this time, does not support SMB3. If you upgrade the kernel on SELKS, then you can use the filebeat service on SELKS.
After you have install the Linux OS, install Suricata(CentOs or Ubuntu):
Copy SELKS Suricata.yml file from SELKS install. Change logging paths to reflect file share within Azure. Mount SMB3 file share. (More details to come)
-Add info about installing Suricata as a probe (pcap or meta data only or both)
-Add how to mount to Azure file share storage and dump all files to the file share.
-Add configuration files