Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[exporter/clickhouse] exporter fails with IO timeout error under load #30175

Closed
MacNale opened this issue Dec 22, 2023 · 7 comments
Closed

[exporter/clickhouse] exporter fails with IO timeout error under load #30175

MacNale opened this issue Dec 22, 2023 · 7 comments

Comments

@MacNale
Copy link

MacNale commented Dec 22, 2023

Component(s)

exporter/clickhouse

What happened?

Description

Using prometeus-receiver to stress test the clickhouse exporter. Clickhouse exporter fails with IO timeout error.

Steps to Reproduce

  1. Use Prometheus receiver or any other receiver (100K metric samples/minute)
  2. Configure clickhouse exporter using the config below

Single node clickhouse running on linux with 16 core and 64 GB memory

Expected Result

  • Load testing result needs to be published.
  • Exporter is able to ingest 100K metric samples per minute

Actual Result

Timeout error

Collector version

0.81.0

Environment information

Environment

OS: (e.g., "Ubuntu 20.04")
CentOS 8

OpenTelemetry Collector configuration

prometheus:
      config:
        scrape_configs:
          - job_name: 'otel-collector-prom-load'
            scrape_interval: 30s
            static_configs:
              - targets: xxxx
processors:
  batch:
    send_batch_size: 20000
    timeout: 30s
    
 exporters:
  clickhouse:
    endpoint: xxxx
    ...
    timeout: 100s
    retry_on_failure:
      enabled: true
      initial_interval: 5s
      max_interval: 30s
      max_elapsed_time: 300s
      
 pipelines:
    metrics:
      receivers: [prometheus, otlp]
      processors: [resource, batch]
      exporters: [clickhouse]

Log output

2023-12-22T03:25:33.335Z        debug   internal/gauge_metrics.go:145   insert gauge metrics fail       
{"kind": "exporter", "data_type": "metrics", "name": "clickhouse", "cost": 99.529687419}
2023-12-22T03:25:33.335Z        error   exporterhelper/queued_retry.go:165      Exporting failed. 
No more retries left. Dropping data.  {"kind": "exporter", "data_type": "metrics", "name": "clickhouse", "error": "max elapsed time expired insert gauge metrics fail:read: read tcp xxxxxxx:9000 i/o timeout", "dropped_items": 200005}
go.opentelemetry.io/collector/exporter/exporterhelper.(*queuedRetrySender).onTemporaryFailure
        go.opentelemetry.io/collector/[email protected]/exporterhelper/queued_retry.go:165
go.opentelemetry.io/collector/exporter/exporterhelper.(*retrySender).send
        go.opentelemetry.io/collector/[email protected]/exporterhelper/queued_retry.go:407
go.opentelemetry.io/collector/exporter/exporterhelper.(*metricsSenderWithObservability).send
        go.opentelemetry.io/collector/[email protected]/exporterhelper/metrics.go:125
go.opentelemetry.io/collector/exporter/exporterhelper.(*queuedRetrySender).start.func1
        go.opentelemetry.io/collector/[email protected]/exporterhelper/queued_retry.go:195
go.opentelemetry.io/collector/exporter/exporterhelper/internal.(*boundedMemoryQueue).StartConsumers.func1
        go.opentelemetry.io/collector/[email protected]/exporterhelper/internal/bounded_memory_queue.go:47

Additional context

No response

@MacNale MacNale added bug Something isn't working needs triage New item requiring triage labels Dec 22, 2023
Copy link
Contributor

Pinging code owners:

See Adding Labels via Comments if you do not have permissions to add labels yourself.

@MacNale MacNale changed the title [clickhouse/exporter] exporter fails with IO timeout error under load [exporter/clickhouse] exporter fails with IO timeout error under load Dec 28, 2023
@crobert-1
Copy link
Member

crobert-1 commented Feb 6, 2024

Hello @MacNale, it looks like from the performance guide the exporter should be able to handle this load.

@hanjm, @dmitryax, @Frapschen: Is there something simple here that we're missing? Have you been able to take a look at this yet?

Copy link
Contributor

github-actions bot commented Apr 8, 2024

This issue has been inactive for 60 days. It will be closed in 60 days if there is no activity. To ping code owners by adding a component label, see Adding Labels via Comments, or if you are unsure of which component this issue relates to, please ping @open-telemetry/collector-contrib-triagers. If this issue is still relevant, please ping the code owners or leave a comment explaining why it is still relevant. Otherwise, please close it.

Pinging code owners:

See Adding Labels via Comments if you do not have permissions to add labels yourself.

@github-actions github-actions bot added the Stale label Apr 8, 2024
@mshustov
Copy link

mshustov commented Apr 8, 2024

@SpencerTorres PTAL

@crobert-1 crobert-1 removed the Stale label Apr 8, 2024
Copy link
Contributor

This issue has been inactive for 60 days. It will be closed in 60 days if there is no activity. To ping code owners by adding a component label, see Adding Labels via Comments, or if you are unsure of which component this issue relates to, please ping @open-telemetry/collector-contrib-triagers. If this issue is still relevant, please ping the code owners or leave a comment explaining why it is still relevant. Otherwise, please close it.

Pinging code owners:

See Adding Labels via Comments if you do not have permissions to add labels yourself.

@github-actions github-actions bot added the Stale label Jun 10, 2024
Copy link
Contributor

github-actions bot commented Aug 9, 2024

This issue has been closed as inactive because it has been stale for 120 days with no activity.

@github-actions github-actions bot closed this as not planned Won't fix, can't repro, duplicate, stale Aug 9, 2024
@SpencerTorres
Copy link
Member

@MacNale let me know if this is still an issue. We've seen great performance from the exporter under load in production deployments. Perhaps this is specific to metrics?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

4 participants