Elasticsearch Recon Ingestion Scripts (ERIS) 🔎
Go to file
Dionysus b018da4e4d
Full source commenting, uniformity in testing function, records stored as a list by default incase an IP address yields multiple PTR records
2024-03-11 19:18:03 -04:00
helpers MassDNS ingestion script now caches the previous record to support IP addresses that yield more than one PTR record (field turned into a list when +1). Records will now upsert so MassDNS can be streaming into ES 24/7= 2024-03-07 21:57:10 -05:00
ingestors Full source commenting, uniformity in testing function, records stored as a list by default incase an IP address yields multiple PTR records 2024-03-11 19:18:03 -04:00
LICENSE Updated README, fixed issue using the wrong domain in records for zone file ingestion (woops) 2024-01-20 10:53:55 -05:00
README.md Fixed issue with ingest_certs and the ingestion function signature. Simple placeholder argument (un-used) added to maintain function uniformity 2024-03-07 23:33:20 -05:00
eris.py Many bugs fixed in sniffer and async model. 2024-03-08 12:13:57 -05:00
sniff_patch.py Many bugs fixed in sniffer and async model. 2024-03-08 12:13:57 -05:00

README.md

Elasticsearch Recon Ingestion Scripts (ERIS)

A utility for ingesting various large scale reconnaissance data logs into Elasticsearch

The is a suite of tools to aid in the ingestion of recon data from various sources (httpx, masscan, zonefiles, etc) into an Elasticsearch cluster. The entire codebase is designed with asynconous processing, aswell as load balancing ingestion across all of the nodes in your cluster. Additionally, live data ingestion is supported from many of the sources supported. This means data can be directly processed and ingested into your Elasticsearch cluster instantly. The structure allows for the developement of "modules" or "plugins" if you will, to quickly create custom ingestion helpers for anything!

Prerequisites

Usage

python eris.py [options] <input>

Note: The <input> can be a file or a directory of files, depending on the ingestion script.

Options

General arguments
Argument Description
input_path Path to the input file or directory
--watch Create or watch a FIFO for real-time indexing
Elasticsearch arguments
Argument Description Default
--host Elasticsearch host http://localhost/
--port Elasticsearch port 9200
--user Elasticsearch username elastic
--password Elasticsearch password $ES_PASSWORD
--api-key Elasticsearch API Key for authentication $ES_APIKEY
--self-signed Elasticsearch connection with a self-signed certificate
Elasticsearch indexing arguments
Argument Description Default
--index Elasticsearch index name Depends on ingestor
--pipeline Use an ingest pipeline for the index
--replicas Number of replicas for the index 1
--shards Number of shards for the index 1
Performance arguments
Argument Description Default
--chunk-max Maximum size in MB of a chunk 100
--chunk-size Number of records to index in a chunk 50000
--retries Number of times to retry indexing a chunk before failing 100
--timeout Number of seconds to wait before retrying a chunk 60
Ingestion arguments
Argument Description
--certs Index Certstream records
--httpx Index HTTPX records
--masscan Index Masscan records
--massdns Index massdns records
--zone Index zone DNS records

This ingestion suite will use the built in node sniffer, so by connecting to a single node, you can load balance across the entire cluster. It is good to know how much nodes you have in the cluster to determine how to fine tune the arguments for the best performance, based on your environment.

Roadmap

  • Create a module for RIR database ingestion (WHOIS, delegations, transfer, ASN mapping, peering, etc)
  • Dynamically update the batch metrics when the sniffer adds or removes nodes.

Mirrors for this repository: acid.vegasSuperNETsGitHubGitLabCodeberg