Elasticsearch Recon Ingestion Scripts (ERIS) 🔎
Go to file
2024-02-02 13:21:59 -05:00
helpers Added an asyncronous sniffer patch in helpers 2024-02-02 13:21:59 -05:00
ingestors Ingestion agents are now modular, FIFO live ingestion added 2024-02-02 00:11:18 -05:00
eris.py Ingestion agents are now modular, FIFO live ingestion added 2024-02-02 00:11:18 -05:00
LICENSE Updated README, fixed issue using the wrong domain in records for zone file ingestion (woops) 2024-01-20 10:53:55 -05:00
README.md Ingestion agents are now modular, FIFO live ingestion added 2024-02-02 00:11:18 -05:00
sniff_patch.py Ingestion agents are now modular, FIFO live ingestion added 2024-02-02 00:11:18 -05:00

Elasticsearch Recon Ingestion Scripts (ERIS)

A utility for ingesting various large scale reconnaissance data logs into Elasticsearch

Work In Progress

Prerequisites

Usage

python eris.py [options] <input>

Note: The <input> can be a file or a directory of files, depending on the ingestion script.

Options

General arguments
Argument Description
input_path Path to the input file or directory
--dry-run Dry run (do not index records to Elasticsearch)
--watch Create or watch a FIFO for real-time indexing
Elasticsearch arguments
Argument Description Default
--host Elasticsearch host localhost
--port Elasticsearch port 9200
--user Elasticsearch username elastic
--password Elasticsearch password $ES_PASSWORD
--api-key Elasticsearch API Key for authentication (format must be api_key:api_secret) $ES_APIKEY
--self-signed Elasticsearch connection with a self-signed certificate
Elasticsearch indexing arguments
Argument Description Default
--index Elasticsearch index name Depends on ingestor
--pipeline Use an ingest pipeline for the index
--replicas Number of replicas for the index 1
--shards Number of shards for the index 1
Performance arguments
Argument Description Default
--chunk-max Maximum size in MB of a chunk 10
--chunk-size Number of records to index in a chunk 5000
--chunk-threads Number of threads to use when indexing in chunks 2
--retries Number of times to retry indexing a chunk before failing 10
--timeout Number of seconds to wait before retrying a chunk 30
Ingestion arguments
Argument Description
--httpx Index HTTPX records
--masscan Index Masscan records
--massdns Index massdns records
--zone Index zone DNS records

Using --batch-threads as 4 and --batch-size as 10000 with 3 nodes would process 120,000 records before indexing 40,000 per node. Take these kind of metrics int account when consider how much records you want to process at once and the memory limitations of your environment, aswell as the networking constraint it may have ono your node(s), depending on the size of your cluster.

Operations

This ingestion suite will use the built in node sniffer, so by connecting to a single node, you can load balance across the entire cluster. It is good to know how much nodes you have in the cluster to determine how to fine tune the arguments for the best performance, based on your environment.

Changelog

  • The --watch feature now uses a FIFO to do live ingestion.
  • Isolated eris.py into it's own file and seperated the ingestion agents into their own modules.

Mirrors for this repository: acid.vegasSuperNETsGitHubGitLabCodeberg