-
Notifications
You must be signed in to change notification settings - Fork 69
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Feature Request] OpenTelemetry Metrics OLTP/HTTP support #647
Comments
That is unfortunate that such a platform doesn't support such a common OTLP approach. This is implemented in our Rust Core layer, so I have opened an issue over there to support HTTP: temporalio/sdk-core#820. Once updated there we will apply here (may just be an |
Sorry temporal, we'll try to get grpc added as soon as possible. |
@slingshotvfx as a workaround, I would recommend deploying an OTEL collector which can receive grpc/http telemetry and export to Logfire via HTTP. This would be a better solution in the long run, as you ideally don't want to send telemetry directly to the backend via OTLP from your app. A collector allows you to add batching, retry/fault tolerance, and transformations to customise the metric labels. Here's a quick example in Docker Compose (not fully tested) using Grafana Alloy (I prefer it over the standard OTel collector for its UI and live debugging): Details
docker-compose.yaml: services:
alloy:
image: grafana/alloy:v1.3.1
environment:
- OTLP_EXPORTER_ENDPOINT="https://log-fire.com/whatever"
ports:
- "4317:4317" # OTLP grpc receiver
- "4318:4318" # OTLP http receiver
- "12345:12345" # Debug UI
volumes:
- alloy_gateway:/var/lib/alloy/data
- type: bind
source: ./config.alloy
target: /etc/alloy/config.alloy
command:
- "run"
- "--server.http.listen-addr=0.0.0.0:12345"
- "--storage.path=/var/lib/alloy/data"
- "--stability.level=experimental" # enables live debug mode
- "/etc/alloy/config.alloy" config.alloy: livedebugging {
// Enable live debugging for the collector
// See http://localhost:12345
enabled = true
}
otelcol.receiver.otlp "default" {
// Setup OTLP receiver to accept metrics, traces, and logs from instrumented services
// https://grafana.com/docs/alloy/latest/reference/components/otelcol/otelcol.receiver.otlp/
grpc {
endpoint = "0.0.0.0:4317"
}
http {
endpoint = "0.0.0.0:4318"
}
output {
metrics = [otelcol.processor.batch.input]
# traces = []
# logs = []
}
}
otelcol.processor.batch "metrics" {
// Adds batching to metrics
// https://grafana.com/docs/alloy/latest/reference/components/otelcol/otelcol.processor.batch/
timeout = "200ms"
output {
metrics = [otelcol.exporter.otlphttp.logfire.input]
}
}
// Can add additional components as needed
otelcol.exporter.otlphttp "logfire" {
// Exports telemetry to another OTLP receiver via http
// https://grafana.com/docs/alloy/latest/reference/components/otelcol/otelcol.exporter.otlphttp/
client {
endpoint = env("OTLP_EXPORTER_ENDPOINT")
// You'll likely need to configure auth / tls to push to your backend
}
} Then you just need to configure your Python app to send telemetry to |
Is your feature request related to a problem? Please describe.
Right now the Python SDK (and possibly other Core SDKs?) only support OLTP over gRPC, but some services (like Pydantic Logfire) require OLTP over HTTPS
When trying to export metrics to these services today, I receive the following error:
OpenTelemetry metrics error occurred. Metrics exporter otlp failed with the grpc server returns error (Internal error): , detailed error message: protocol error: received message with invalid compression flag: 60 (valid flags are 0 and 1) while receiving response with status: 403 Forbidden
Describe the solution you'd like
Support for exporting OTel Metrics to services via OLTP/HTTP.
Additional context
This would bring the Python SDK up to parity with the Go SDK and possibly others which do support HTTP.
The text was updated successfully, but these errors were encountered: