From c1ecc0afddc87ced2aac6adf9d4d8efc35858126 Mon Sep 17 00:00:00 2001 From: Antoine Toulme Date: Tue, 6 Feb 2024 12:55:07 -0800 Subject: [PATCH] Remove outdated examples (#4268) --- examples/docker-compose/.env | 3 - examples/docker-compose/docker-compose.yml | 17 - .../splunk-otel-collector-agent.yaml | 223 --------- .../splunk-otel-collector-gateway.yaml | 258 ----------- examples/nomad/README.md | 12 - examples/nomad/otel-demo.nomad | 434 ------------------ .../otel-logs-internal-monitoring/.gitignore | 1 - .../otel-logs-internal-monitoring/README.md | 47 -- .../docker-compose.yml | 56 --- .../otel-collector-config.yml | 46 -- .../prometheus.yml | 7 - .../otel-logs-internal-monitoring/splunk.yml | 11 - 12 files changed, 1115 deletions(-) delete mode 100644 examples/docker-compose/.env delete mode 100644 examples/docker-compose/docker-compose.yml delete mode 100644 examples/kubernetes-yaml/splunk-otel-collector-agent.yaml delete mode 100644 examples/kubernetes-yaml/splunk-otel-collector-gateway.yaml delete mode 100644 examples/nomad/README.md delete mode 100644 examples/nomad/otel-demo.nomad delete mode 100644 examples/otel-logs-internal-monitoring/.gitignore delete mode 100644 examples/otel-logs-internal-monitoring/README.md delete mode 100644 examples/otel-logs-internal-monitoring/docker-compose.yml delete mode 100644 examples/otel-logs-internal-monitoring/otel-collector-config.yml delete mode 100644 examples/otel-logs-internal-monitoring/prometheus.yml delete mode 100644 examples/otel-logs-internal-monitoring/splunk.yml diff --git a/examples/docker-compose/.env b/examples/docker-compose/.env deleted file mode 100644 index a8d72b32c1..0000000000 --- a/examples/docker-compose/.env +++ /dev/null @@ -1,3 +0,0 @@ -SPLUNK_ACCESS_TOKEN=12345 -SPLUNK_MEMORY_TOTAL_MIB=1024 -SPLUNK_REALM=us0 diff --git a/examples/docker-compose/docker-compose.yml b/examples/docker-compose/docker-compose.yml deleted file mode 100644 index 1d06d00c34..0000000000 --- a/examples/docker-compose/docker-compose.yml +++ /dev/null @@ -1,17 +0,0 @@ -version: "3" -services: - splunk-otel-collector: - image: quay.io/signalfx/splunk-otel-collector:0.59.1 - environment: - - SPLUNK_ACCESS_TOKEN=${SPLUNK_ACCESS_TOKEN} - - SPLUNK_REALM=${SPLUNK_REALM} - ports: - - "13133:13133" - - "14250:14250" - - "14268:14268" - - "4317:4317" - - "6060:6060" - - "8888:8888" - - "9080:9080" - - "9411:9411" - - "9943:9943" diff --git a/examples/kubernetes-yaml/splunk-otel-collector-agent.yaml b/examples/kubernetes-yaml/splunk-otel-collector-agent.yaml deleted file mode 100644 index 98840b2a07..0000000000 --- a/examples/kubernetes-yaml/splunk-otel-collector-agent.yaml +++ /dev/null @@ -1,223 +0,0 @@ ---- -apiVersion: v1 -kind: ConfigMap -metadata: - name: otel-collector-conf - labels: - app: opentelemetry - component: otel-collector-conf -data: - otel-collector-config: | - extensions: - health_check: - endpoint: 0.0.0.0:13133 - http_forwarder: - ingress: - endpoint: 0.0.0.0:6060 - egress: - # TODO: Ensure this is set properly - endpoint: "https://api.${SPLUNK_REALM}.signalfx.com" - zpages: - memory_ballast: - # In general, the ballast should be set to 1/3 of the collector's memory, the limit - # should be 90% of the collector's memory. - # The simplest way to specify the ballast size is set the value of SPLUNK_BALLAST_SIZE_MIB env variable. - # TODO: Ensure this is set properly - size_mib: ${SPLUNK_BALLAST_SIZE_MIB} - receivers: - hostmetrics: - collection_interval: 10s - scrapers: - cpu: - disk: - filesystem: - memory: - network: - # System load average metrics https://en.wikipedia.org/wiki/Load_(computing) - load: - # Paging/Swap space utilization and I/O metrics - paging: - # Aggregated system process count metrics - processes: - # System processes metrics, disabled by default - # process: - jaeger: - protocols: - grpc: - endpoint: 0.0.0.0:14250 - thrift_binary: - endpoint: 0.0.0.0:6832 - thrift_compact: - endpoint: 0.0.0.0:6831 - thrift_http: - endpoint: 0.0.0.0:14268 - otlp: - protocols: - grpc: - endpoint: 0.0.0.0:4317 - http: - endpoint: 0.0.0.0:4318 - sapm: - endpoint: 0.0.0.0:7276 - signalfx: - endpoint: 0.0.0.0:9943 - # This section is used to collect OpenTelemetry metrics - # Even if just a SignalFx µAPM customer, these metrics are included - prometheus: - config: - scrape_configs: - - job_name: 'otel-collector' - scrape_interval: 10s - static_configs: - - targets: ['localhost:8888'] - # If you want to use the environment filter - # In the SignalFx dashboard - #labels: - #environment: demo - metric_relabel_configs: - - source_labels: [ __name__ ] - regex: '.*grpc_io.*' - action: drop - # Enable Zipkin to support Istio Mixer Adapter - # https://github.com/signalfx/signalfx-istio-adapter - zipkin: - endpoint: 0.0.0.0:9411 - processors: - batch: - # Optional: If you have a different environment tag name - # If this option is enabled it must be added to the pipeline section below - #attributes/copyfromexistingkey: - #actions: - #- key: environment - #from_attribute: YOUR_EXISTING_TAG_NAME - #action: upsert - # Optional: If you want to add an environment tag - # If this option is enabled it must be added to the pipeline section below - #attributes/newenvironment: - #actions: - #- key: environment - #value: "YOUR_ENVIRONMENT_NAME" - #action: insert - # Enabling the memory_limiter is strongly recommended for every pipeline. - # Configuration is based on the amount of memory allocated to the collector. - # The configuration below assumes 2GB of memory. In general, the ballast - # should be set to 1/3 of the collector's memory, the limit should be 90% of - # the collector's memory, and the spike should be 25% of the - # collector's memory. In addition, the "--mem-ballast-size-mib" CLI - # flag must be set to the same value as the "ballast_size_mib". For more - # information, see - # https://github.com/open-telemetry/opentelemetry-collector/blob/main/processor/memorylimiter/README.md - memory_limiter: - # TODO: Ensure this is set properly - ballast_size_mib: 168 - check_interval: 2s - # TODO: Ensure this is set properly - limit_mib: 460 - exporters: - # Traces - sapm: - # TODO: Ensure this is set properly - access_token: "${SPLUNK_ACCESS_TOKEN}" - # TODO: Ensure this is set properly - endpoint: "https://ingest.${SPLUNK_REALM}.signalfx.com/v2/trace" - # Metrics - signalfx: - # TODO: Ensure this is set properly - access_token: "${SPLUNK_ACCESS_TOKEN}" - # TODO: Ensure this is set properly - realm: "${SPLUNK_REALM}" - sync_host_metadata: true - service: - extensions: [health_check, http_forwarder, zpages] - pipelines: - traces: - receivers: [jaeger, otlp, sapm, zipkin] - processors: [memory_limiter, batch] - exporters: [sapm, signalfx] - metrics: - receivers: [otlp, signalfx, prometheus, hostmetrics] - processors: [memory_limiter, batch] - exporters: [signalfx] - logs: - receivers: [signalfx] - processors: [memory_limiter, batch] - exporters: [signalfx] ---- -apiVersion: apps/v1 -kind: DaemonSet -metadata: - name: otel-agent - labels: - app: opentelemetry - component: otel-agent -spec: - selector: - matchLabels: - app: opentelemetry - component: otel-agent - minReadySeconds: 5 - progressDeadlineSeconds: 120 - template: - metadata: - labels: - app: opentelemetry - component: otel-agent - spec: - containers: - - command: - - "/otelcol" - - "--config=/conf/otel-collector-config.yaml" - # TODO: Ensure this is set properly - # WARN: It is not recommended to use the latest tag in production deployments! - image: quay.io/signalfx/splunk-otel-collector:latest - name: otel-agent - resources: - limits: - # TODO: Ensure this is set properly - cpu: 500m - # TODO: Ensure this is set properly - memory: 512Mi - requests: - cpu: 100m - memory: 100Mi - ports: - - containerPort: 13133 # Default endpoint for Health Check. - - containerPort: 14250 # Default endpoint for Jaeger gRPC receiver. - - containerPort: 14268 # Default endpoint for Jaeger Thrift receiver. - - containerPort: 4317 # Default endpoint for OpenTelemetry gRPC receiver. - - containerPort: 6060 # Default endpoint for HTTP Forwarder extension. - - containerPort: 6831 # Default endpoint for Jaeger Thrift compact receiver. - protocol: UDP - - containerPort: 6832 # Default endpoint for Jaeger Thrift binary receiver. - protocol: UDP - - containerPort: 7276 # Default endpoint for SignalFx APM receiver. - - containerPort: 8888 # Default endpoint for querying metrics. - - containerPort: 9411 # Default endpoint for Zipkin trace receiver. - - containerPort: 9943 # Default endpoint for SignalFx metrics receiver. - volumeMounts: - - name: otel-collector-config-vol - mountPath: /conf -# - name: otel-collector-secrets -# mountPath: /secrets - livenessProbe: - httpGet: - path: / - port: 13133 # Health Check extension default port. - readinessProbe: - httpGet: - path: / - port: 13133 # Health Check extension default port. - volumes: - - configMap: - name: otel-collector-conf - items: - - key: otel-collector-config - path: otel-collector-config.yaml - name: otel-collector-config-vol -# - secret: -# name: otel-collector-secrets -# items: -# - key: cert.pem -# path: cert.pem -# - key: key.pem -# path: key.pem diff --git a/examples/kubernetes-yaml/splunk-otel-collector-gateway.yaml b/examples/kubernetes-yaml/splunk-otel-collector-gateway.yaml deleted file mode 100644 index 8610af3524..0000000000 --- a/examples/kubernetes-yaml/splunk-otel-collector-gateway.yaml +++ /dev/null @@ -1,258 +0,0 @@ ---- -apiVersion: v1 -kind: ConfigMap -metadata: - name: otel-collector-conf - labels: - app: opentelemetry - component: otel-collector-conf -data: - otel-collector-config: | - extensions: - health_check: - endpoint: 0.0.0.0:13133 - http_forwarder: - ingress: - endpoint: 0.0.0.0:6060 - egress: - # TODO: Ensure this is set properly - endpoint: "https://api.${SPLUNK_REALM}.signalfx.com" - zpages: - memory_ballast: - # In general, the ballast should be set to 1/3 of the collector's memory, the limit - # should be 90% of the collector's memory. - # The simplest way to specify the ballast size is set the value of SPLUNK_BALLAST_SIZE_MIB env variable. - # TODO: Ensure this is set properly - size_mib: ${SPLUNK_BALLAST_SIZE_MIB} - receivers: - hostmetrics: - collection_interval: 10s - scrapers: - cpu: - disk: - filesystem: - memory: - network: - # System load average metrics https://en.wikipedia.org/wiki/Load_(computing) - load: - # Paging/Swap space utilization and I/O metrics - paging: - # Aggregated system process count metrics - processes: - # System processes metrics, disabled by default - # process: - jaeger: - protocols: - grpc: - endpoint: 0.0.0.0:14250 - thrift_binary: - endpoint: 0.0.0.0:6832 - thrift_compact: - endpoint: 0.0.0.0:6831 - thrift_http: - endpoint: 0.0.0.0:14268 - otlp: - protocols: - grpc: - endpoint: 0.0.0.0:4317 - http: - endpoint: 0.0.0.0:4318 - sapm: - endpoint: 0.0.0.0:7276 - signalfx: - endpoint: 0.0.0.0:9943 - # This section is used to collect OpenTelemetry metrics - # Even if just a SignalFx µAPM customer, these metrics are included - prometheus: - config: - scrape_configs: - - job_name: 'otel-collector' - scrape_interval: 10s - static_configs: - - targets: ['localhost:8888'] - # If you want to use the environment filter - # In the SignalFx dashboard - #labels: - #environment: demo - metric_relabel_configs: - - source_labels: [ __name__ ] - regex: '.*grpc_io.*' - action: drop - # Enable Zipkin to support Istio Mixer Adapter - # https://github.com/signalfx/signalfx-istio-adapter - zipkin: - endpoint: 0.0.0.0:9411 - processors: - batch: - # Optional: If you have a different environment tag name - # If this option is enabled it must be added to the pipeline section below - #attributes/copyfromexistingkey: - #actions: - #- key: environment - #from_attribute: YOUR_EXISTING_TAG_NAME - #action: upsert - # Optional: If you want to add an environment tag - # If this option is enabled it must be added to the pipeline section below - #attributes/newenvironment: - #actions: - #- key: environment - #value: "YOUR_ENVIRONMENT_NAME" - #action: insert - # Enabling the memory_limiter is strongly recommended for every pipeline. - # Configuration is based on the amount of memory allocated to the collector. - # The configuration below assumes 2GB of memory. In general, the ballast - # should be set to 1/3 of the collector's memory, the limit should be 90% of - # the collector's memory, and the spike should be 25% of the - # collector's memory. In addition, the "--mem-ballast-size-mib" CLI - # flag must be set to the same value as the "ballast_size_mib". For more - # information, see - # https://github.com/open-telemetry/opentelemetry-collector/blob/main/processor/memorylimiter/README.md - memory_limiter: - # TODO: Ensure this is set properly - ballast_size_mib: 168 - check_interval: 2s - # TODO: Ensure this is set properly - limit_mib: 460 - exporters: - # Traces - sapm: - # TODO: Ensure this is set properly - access_token: "${SPLUNK_ACCESS_TOKEN}" - # TODO: Ensure this is set properly - endpoint: "https://ingest.${SPLUNK_REALM}.signalfx.com/v2/trace" - # Metrics - signalfx: - # TODO: Ensure this is set properly - access_token: "${SPLUNK_ACCESS_TOKEN}" - # TODO: Ensure this is set properly - realm: "${SPLUNK_REALM}" - service: - extensions: [health_check, http_forwarder, zpages] - pipelines: - traces: - receivers: [jaeger, otlp, sapm, zipkin] - processors: [memory_limiter, batch] - exporters: [sapm, signalfx] - metrics: - receivers: [otlp, signalfx, prometheus, hostmetrics] - processors: [memory_limiter, batch] - exporters: [signalfx] - logs: - receivers: [signalfx] - processors: [memory_limiter, batch] - exporters: [signalfx] ---- -apiVersion: v1 -kind: Service -metadata: - name: otel-collector - labels: - app: opentelemetry - component: otel-collector -spec: - ports: - - name: health # Default Health Check port. - port: 13133 - - name: jaeger-grpc # Default endpoint for Jaeger gRPC receiver. - port: 14250 - - name: thrift-binary # Default endpoint for Jaeger Thrift receiver. - port: 6832 - protocol: UDP - - name: thrift-compact - port: 6831 - protocol: UDP - - name: thrift-http # Default endpoint for Jaeger Thrift receiver. - port: 14268 - - name: otlp # Default endpoint for OTLP gRPC receiver. - port: 4317 - - name: otlp-http # Default endpoint for OTLP HTTP receiver. - port: 4318 - - name: otlp-http-old # Legacy endpoint for OTLP HTTP receiver. - port: 55681 - - name: http-forwarder # Default endpoint for HTTP Forwarder extension. - port: 6060 - - name: sapm # Default endpoint for SignalFx APM receiver. - port: 7276 - - name: zipkin # Default endpoint for Zipkin trace receiver. - port: 9411 - - name: signalfx # Default endpoint for SignalFx metrics receiver. - port: 9943 - selector: - component: otel-collector ---- -apiVersion: apps/v1 -kind: Deployment -metadata: - name: otel-collector - labels: - app: opentelemetry - component: otel-collector -spec: - selector: - matchLabels: - app: opentelemetry - component: otel-collector - minReadySeconds: 5 - progressDeadlineSeconds: 120 - # TODO: Ensure this is set properly - replicas: 1 - template: - metadata: - labels: - app: opentelemetry - component: otel-collector - spec: - containers: - - command: - - "/otelcol" - - "--config=/conf/otel-collector-config.yaml" - # TODO: Ensure this is set properly - # WARN: It is not recommended to use the latest tag in production deployments! - image: quay.io/signalfx/splunk-otel-collector:latest - name: otel-collector - resources: - limits: - # TODO: Ensure this is set properly - cpu: 500m - # TODO: Ensure this is set properly - memory: 512Mi - requests: - cpu: 200m - memory: 400Mi - ports: - - containerPort: 13133 # Default endpoint for Health Check. - - containerPort: 14250 # Default endpoint for Jaeger gRPC receiver. - - containerPort: 14268 # Default endpoint for Jaeger Thrift receiver. - - containerPort: 4317 # Default endpoint for OpenTelemetry gRPC receiver. - - containerPort: 6060 # Default endpoint for HTTP Forwarder extension. - - containerPort: 7276 # Default endpoint for SignalFx APM receiver. - - containerPort: 8888 # Default endpoint for querying metrics. - - containerPort: 9411 # Default endpoint for Zipkin trace receiver. - - containerPort: 9943 # Default endpoint for SignalFx metrics receiver. - volumeMounts: - - name: otel-collector-config-vol - mountPath: /conf -# - name: otel-collector-secrets -# mountPath: /secrets - livenessProbe: - httpGet: - path: / - port: 13133 # Health Check extension default port. - readinessProbe: - httpGet: - path: / - port: 13133 # Health Check extension default port. - volumes: - - configMap: - name: otel-collector-conf - items: - - key: otel-collector-config - path: otel-collector-config.yaml - name: otel-collector-config-vol -# - secret: -# name: otel-collector-secrets -# items: -# - key: cert.pem -# path: cert.pem -# - key: key.pem -# path: key.pem diff --git a/examples/nomad/README.md b/examples/nomad/README.md deleted file mode 100644 index 5b830ae3f7..0000000000 --- a/examples/nomad/README.md +++ /dev/null @@ -1,12 +0,0 @@ -# Nomad demo Example - -This example showcases how the collector can send data to a Splunk Enterprise deployment over the Nomad cluster. - -The example will deploy the Splunk OpenTelemetry Collector job as `agent` and `gateway` to collect metrics and traces and export them using the `SignalFx` exporter. This will also deploy `load generators` job to generate traces of `Zipkin` and `Jaeger`. - -To deploy the example, check out this git repository, open a terminal and in this directory type: -```bash -$ git clone https://github.com/signalfx/splunk-otel-collector.git -$ cd splunk-otel-collector/examples/nomad -$ nomad run otel-demo.nomad -``` diff --git a/examples/nomad/otel-demo.nomad b/examples/nomad/otel-demo.nomad deleted file mode 100644 index a04f0d9bea..0000000000 --- a/examples/nomad/otel-demo.nomad +++ /dev/null @@ -1,434 +0,0 @@ -job "otel-demo" { - datacenters = ["dc1"] - type = "service" - - # Collector - group "otel-gateway" { - count = 1 - - network { - port "metrics" { - to = 8889 - } - - # Receivers - port "otlp" { - to = 4317 - } - - port "jaeger-grpc" { - to = 14250 - } - - port "jaeger-thrift-http" { - to = 14268 - } - - port "zipkin" { - to = 9411 - } - - port "signalfx" { - to = 9943 - } - - # Extensions - port "health-check" { - to = 13133 - } - - port "zpages" { - to = 55679 - } - } - - service { - name = "otel-gateway" - port = "health-check" - tags = ["health"] - - check { - type = "http" - port = "health-check" - path = "/" - interval = "5s" - timeout = "2s" - } - } - - service { - name = "otel-gateway" - port = "otlp" - tags = ["otlp"] - } - - service { - name = "otel-gateway" - port = "jaeger-grpc" - tags = ["jaeger-grpc"] - } - - service { - name = "otel-gateway" - port = "jaeger-thrift-http" - tags = ["jaeger-thrift-http"] - } - - service { - name = "otel-gateway" - port = "zipkin" - tags = ["zipkin"] - } - - service { - name = "otel-gateway" - port = "signalfx" - tags = ["signalfx"] - } - - service { - name = "otel-gateway" - port = "metrics" - tags = ["metrics"] - } - - service { - name = "otel-gateway" - port = "zpages" - tags = ["zpages"] - } - - task "otel-gateway" { - driver = "docker" - - config { - image = "quay.io/signalfx/splunk-otel-collector:latest" - force_pull = false - entrypoint = [ - "/otelcol", - "--config=local/config/otel-gateway-config.yaml", - "--metrics-addr=0.0.0.0:8889", - ] - - ports = [ - "metrics", - "otlp", - "jaeger-grpc", - "jaeger-thrift-http", - "zipkin", - "health-check", - "zpages", - "signalfx", - ] - } - - env { - SPLUNK_ACCESS_TOKEN = "" - SPLUNK_REALM = "" - SPLUNK_MEMORY_TOTAL_MIB = 500 - } - - resources { - cpu = 500 - memory = 500 - } - - template { - data = < docker-compose up -``` - -Splunk will become available on port 18000. You can login on [http://localhost:18000](http://localhost:18000) with `admin` and `changeme`. - -Once logged in, visit the [search application](http://localhost:18000/en-US/app/search) to see the logs collected by Splunk. - -Visit [http://localhost:8888](http://localhost:8888) to see the metrics emitted by the OpenTelemetry Collector. - -Here is a sample excerpt of metrics: -``` -# HELP otelcol_exporter_enqueue_failed_log_records Number of log records failed to be added to the sending queue. -# TYPE otelcol_exporter_enqueue_failed_log_records counter -otelcol_exporter_enqueue_failed_log_records{exporter="splunk_hec/logs",service_instance_id="a266e329-b02f-459f-9ec7-2e06021d4135",service_version="v0.57.0"} 0 -# HELP otelcol_exporter_enqueue_failed_metric_points Number of metric points failed to be added to the sending queue. -# TYPE otelcol_exporter_enqueue_failed_metric_points counter -otelcol_exporter_enqueue_failed_metric_points{exporter="splunk_hec/logs",service_instance_id="a266e329-b02f-459f-9ec7-2e06021d4135",service_version="v0.57.0"} 0 -# HELP otelcol_exporter_enqueue_failed_spans Number of spans failed to be added to the sending queue. -# TYPE otelcol_exporter_enqueue_failed_spans counter -otelcol_exporter_enqueue_failed_spans{exporter="splunk_hec/logs",service_instance_id="a266e329-b02f-459f-9ec7-2e06021d4135",service_version="v0.57.0"} 0 -# HELP otelcol_exporter_queue_capacity Fixed capacity of the retry queue (in batches) -# TYPE otelcol_exporter_queue_capacity gauge -otelcol_exporter_queue_capacity{exporter="splunk_hec/logs",service_instance_id="a266e329-b02f-459f-9ec7-2e06021d4135",service_version="v0.57.0"} 5000 -# HELP otelcol_exporter_queue_size Current size of the retry queue (in batches) -# TYPE otelcol_exporter_queue_size gauge -otelcol_exporter_queue_size{exporter="splunk_hec/logs",service_instance_id="a266e329-b02f-459f-9ec7-2e06021d4135",service_version="v0.57.0"} 0 -# HELP otelcol_exporter_sent_log_records Number of log record successfully sent to destination. -# TYPE otelcol_exporter_sent_log_records counter -otelcol_exporter_sent_log_records{exporter="splunk_hec/logs",service_instance_id="a266e329-b02f-459f-9ec7-2e06021d4135",service_version="v0.57.0"} 45 -``` - -Additionally, this example runs Prometheus on port 9090. Visit [http://localhost:9090](http://localhost:9090), -and enter a query to examine metrics exposed by the OpenTelemetry Collector. - -For more information on the metrics exposed by the collector, read on to the [official OpenTelemetry monitoring documentation](https://github.com/open-telemetry/opentelemetry-collector/blob/main/docs/monitoring.md). - -For more information on troubleshooting the collector, please refer to the [official OpenTelemetry documentation](https://github.com/open-telemetry/opentelemetry-collector/blob/main/docs/troubleshooting.md). \ No newline at end of file diff --git a/examples/otel-logs-internal-monitoring/docker-compose.yml b/examples/otel-logs-internal-monitoring/docker-compose.yml deleted file mode 100644 index 51b92e0d6a..0000000000 --- a/examples/otel-logs-internal-monitoring/docker-compose.yml +++ /dev/null @@ -1,56 +0,0 @@ -version: "3" -services: - # Bash script producing logs. - logging: - image: bash:latest - container_name: logging - command: 'bash -c "while(true) do echo \"$$(date) new message\" >> /output/file.log ; sleep 1; done"' - volumes: - - ./output:/output - depends_on: - - prometheus - # Splunk Enterprise server: - splunk: - image: splunk/splunk:latest - container_name: splunk - environment: - - SPLUNK_START_ARGS=--accept-license - - SPLUNK_HEC_TOKEN=00000000-0000-0000-0000-0000000000000 - - SPLUNK_PASSWORD=changeme - ports: - - "18000:8000" - healthcheck: - test: ['CMD', 'curl', '-f', 'http://localhost:8000'] - interval: 5s - timeout: 5s - retries: 20 - volumes: - - ./splunk.yml:/tmp/defaults/default.yml - - /opt/splunk/var - - /opt/splunk/etc - # OpenTelemetry Collector - otelcollector: - image: quay.io/signalfx/splunk-otel-collector:0.59.1 - container_name: otelcollector - command: ["--config=/etc/otel-collector-config.yml"] - volumes: - - ./otel-collector-config.yml:/etc/otel-collector-config.yml - - ./output:/output - depends_on: - splunk: - condition: service_healthy - ports: - - "18088:8088" - - "8888:8888" - # Prometheus - prometheus: - image: prom/prometheus:v2.36.2 - container_name: prometheus - volumes: - - ./prometheus.yml:/etc/prometheus/prometheus.yml - command: - - '--config.file=/etc/prometheus/prometheus.yml' - ports: - - "9090:9090" - depends_on: - - otelcollector diff --git a/examples/otel-logs-internal-monitoring/otel-collector-config.yml b/examples/otel-logs-internal-monitoring/otel-collector-config.yml deleted file mode 100644 index 4f76420317..0000000000 --- a/examples/otel-logs-internal-monitoring/otel-collector-config.yml +++ /dev/null @@ -1,46 +0,0 @@ -receivers: - filelog: - include: [ /output/file.log ] - -exporters: - splunk_hec/logs: - # Splunk HTTP Event Collector token. - token: "00000000-0000-0000-0000-0000000000000" - # URL to a Splunk instance to send data to. - endpoint: "https://splunk:8088/services/collector" - # Optional Splunk source: https://docs.splunk.com/Splexicon:Source - source: "output" - # Splunk index, optional name of the Splunk index targeted. - index: "logs" - # Maximum HTTP connections to use simultaneously when sending data. Defaults to 100. - max_connections: 20 - # Whether to disable gzip compression over HTTP. Defaults to false. - disable_compression: false - # HTTP timeout when sending data. Defaults to 10s. - timeout: 10s - tls: - # Whether to skip checking the certificate of the HEC endpoint when sending data over HTTPS. Defaults to false. - # For this demo, we use a self-signed certificate on the Splunk docker instance, so this flag is set to true. - insecure_skip_verify: true - -processors: - batch: - -extensions: - health_check: - endpoint: 0.0.0.0:13133 - pprof: - endpoint: :1888 - zpages: - endpoint: :55679 - -service: - extensions: [pprof, zpages, health_check] - telemetry: - metrics: - address: ":8888" - pipelines: - logs: - receivers: [filelog] - processors: [batch] - exporters: [splunk_hec/logs] diff --git a/examples/otel-logs-internal-monitoring/prometheus.yml b/examples/otel-logs-internal-monitoring/prometheus.yml deleted file mode 100644 index ed84aac379..0000000000 --- a/examples/otel-logs-internal-monitoring/prometheus.yml +++ /dev/null @@ -1,7 +0,0 @@ -global: - scrape_interval: 10s -scrape_configs: - - job_name: otel - static_configs: - - targets: - - otelcollector:8888 \ No newline at end of file diff --git a/examples/otel-logs-internal-monitoring/splunk.yml b/examples/otel-logs-internal-monitoring/splunk.yml deleted file mode 100644 index a7c67c1244..0000000000 --- a/examples/otel-logs-internal-monitoring/splunk.yml +++ /dev/null @@ -1,11 +0,0 @@ -splunk: - conf: - indexes: - directory: /opt/splunk/etc/apps/search/local - content: - logs: - coldPath: $SPLUNK_DB/logs/colddb - datatype: event - homePath: $SPLUNK_DB/logs/db - maxTotalDataSizeMB: 512000 - thawedPath: $SPLUNK_DB/logs/thaweddb