diff --git a/docker-compose.yml b/docker-compose.yml deleted file mode 100644 index d288424..0000000 --- a/docker-compose.yml +++ /dev/null @@ -1,21 +0,0 @@ -version: "3" -services: - logstash: - image: docker.elastic.co/logstash/logstash-oss:6.1.1 - ports: - - "9600:9600" - volumes: - - ./docker-compose/logstash.conf:/usr/share/logstash/pipeline/logstash.conf - - ./docker-compose/events.json:/events.json - command: ["-r"] - exporter: - build: . - ports: - - "9304:9304" - command: ["-logstash.host", "logstash"] - prometheus: - image: prom/prometheus - ports: - - "9090:9090" - volumes: - - ./docker-compose/prometheus.yml:/etc/prometheus/prometheus.yml diff --git a/docker-compose/events.json b/docker-compose/events.json deleted file mode 100644 index 2fad592..0000000 --- a/docker-compose/events.json +++ /dev/null @@ -1,6 +0,0 @@ -{ - "event": { - "else": 1, - "something": "data" - } -} \ No newline at end of file diff --git a/docker-compose/logstash.conf b/docker-compose/logstash.conf deleted file mode 100644 index a534c6c..0000000 --- a/docker-compose/logstash.conf +++ /dev/null @@ -1,22 +0,0 @@ -input { - file { - path => ["/events.json"] - sincedb_path => "/dev/null" - start_position => "beginning" - codec => "json_lines" - } -} - -filter { - if [event][else] { - grok { - match => ["[event][something]", "(?.*)"] - } - } -} - -output { - stdout { - codec => "rubydebug" - } -} \ No newline at end of file diff --git a/docker-compose/prometheus.yml b/docker-compose/prometheus.yml deleted file mode 100644 index b0e8365..0000000 --- a/docker-compose/prometheus.yml +++ /dev/null @@ -1,33 +0,0 @@ -# my global config -global: - scrape_interval: 15s # Set the scrape interval to every 15 seconds. Default is every 1 minute. - evaluation_interval: 15s # Evaluate rules every 15 seconds. The default is every 1 minute. - # scrape_timeout is set to the global default (10s). - -# Alertmanager configuration -alerting: - alertmanagers: - - static_configs: - - targets: - # - alertmanager:9093 - -# Load rules once and periodically evaluate them according to the global 'evaluation_interval'. -rule_files: - # - "first_rules.yml" - # - "second_rules.yml" - -# A scrape configuration containing exactly one endpoint to scrape: -# Here it's Prometheus itself. -scrape_configs: - # The job name is added as a label `job=` to any timeseries scraped from this config. - - job_name: 'prometheus' - - # metrics_path defaults to '/metrics' - # scheme defaults to 'http'. - - static_configs: - - targets: ['localhost:9090'] - - - job_name: 'logstash' - static_configs: - - targets: ['exporter:9304']