Skip to content
/ eris Public

Elasticsearch Recon Ingestion Scripts (ERIS) 🔎

License

Notifications You must be signed in to change notification settings

acidvegas/eris

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

59 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Elasticsearch Recon Ingestion Scripts (ERIS)

A utility for ingesting various large scale reconnaissance data logs into Elasticsearch

The is a suite of tools to aid in the ingestion of recon data from various sources (httpx, masscan, zonefiles, etc) into an Elasticsearch cluster. The entire codebase is designed with asynconous processing, aswell as load balancing ingestion across all of the nodes in your cluster. Additionally, live data ingestion is supported from many of the sources supported. This means data can be directly processed and ingested into your Elasticsearch cluster instantly. The structure allows for the developement of "modules" or "plugins" if you will, to quickly create custom ingestion helpers for anything!

Prerequisites

Usage

python eris.py [options] <input>

Note: The <input> can be a file or a directory of files, depending on the ingestion script.

Options

General arguments
Argument Description
input_path Path to the input file or directory
--watch Create or watch a FIFO for real-time indexing
--log Logging level for file (debug, info, warning, error, critical)

Note: File logging is disabled by default. When enabled, it will log using the Elastic Common Schema (ECS).

Elasticsearch arguments
Argument Description Default
--host Elasticsearch host http://localhost
--port Elasticsearch port 9200
--user Elasticsearch username elastic
--password Elasticsearch password $ES_PASSWORD
--api-key Elasticsearch API Key for authentication $ES_APIKEY
--self-signed Elasticsearch connection with a self-signed certificate
Elasticsearch indexing arguments
Argument Description Default
--index Elasticsearch index name Depends on ingestor
--pipeline Use an ingest pipeline for the index
--replicas Number of replicas for the index 1
--shards Number of shards for the index 1
Performance arguments
Argument Description Default
--chunk-max Maximum size in MB of a chunk 100
--chunk-size Number of records to index in a chunk 50000
--retries Number of times to retry indexing a chunk before failing 100
--timeout Number of seconds to wait before retrying a chunk 60
Ingestion arguments
Argument Description
--certstrem Index Certstream records
--httpx Index HTTPX records
--masscan Index Masscan records
--massdns Index massdns records
--zone Index zone DNS records

This ingestion suite will use the built in node sniffer, so by connecting to a single node, you can load balance across the entire cluster.

Note: The sniffer has been disabled for now due an issue with the 8.x elasticsearch client. The auth headers are not properly sent when enabling the sniffer. A working patch was shared and has been mostly converted in helpers/sniff_patch.py for the async client.

Roadmap

  • Create a module for RIR database ingestion (WHOIS, delegations, transfer, ASN mapping, peering, etc)
  • Dynamically update the batch metrics when the sniffer adds or removes nodes.
  • Fix issue with leftover FIFO files (catch SIGTERM / SIGINT signals)
  • Create a working patch for the async client to properly send auth headers.

Mirrors for this repository: acid.vegas • SuperNETs • GitHub • GitLab • Codeberg