I am trying to deploy Grafana Tempo with below flow:
Synthetic Load generator app ----> Grafana Agent -----> Grafana Tempo ----> Grafana Dashboard
The protocol being used is otlp on HTTP. I am not able to figure out if the traces are being forwarded correctly. The app logs (synthetic load generator) shows that the traces are being emitted but I cannot see them on the grafana dashboard.
Below are the logs from grafana agent::
ts=2022-06-10T20:49:29.150606409Z caller=server.go:195 level=info msg="server listening on addresses" http=127.0.0.1:8080 grpc=127.0.0.1:12346 http_tls_enabled=false grpc_tls_enabled=false
ts=2022-06-10T20:49:29.151289286Z caller=node.go:85 level=info agent=prometheus component=cluster msg="applying config"
ts=2022-06-10T20:49:29Z level=info caller=traces/traces.go:143 msg="Traces Logger Initialized" component=traces
ts=2022-06-10T20:49:29.152656812Z caller=remote.go:180 level=info agent=prometheus component=cluster msg="not watching the KV, none set"
ts=2022-06-10T20:49:29Z level=info caller=traces/instance.go:141 msg="shutting down receiver" component=traces traces_config=default
ts=2022-06-10T20:49:29Z level=info caller=traces/instance.go:141 msg="shutting down processors" component=traces traces_config=default
ts=2022-06-10T20:49:29Z level=info caller=traces/instance.go:141 msg="shutting down exporters" component=traces traces_config=default
ts=2022-06-10T20:49:29Z level=info caller=traces/instance.go:141 msg="shutting down extensions" component=traces traces_config=default
ts=2022-06-10T20:49:29Z level=info caller=builder/exporters_builder.go:255 msg="Exporter was built." component=traces traces_config=default kind=exporter name=otlp/0
ts=2022-06-10T20:49:29Z level=info caller=builder/exporters_builder.go:40 msg="Exporter is starting..." component=traces traces_config=default kind=exporter name=otlp/0
ts=2022-06-10T20:49:29Z level=info caller=builder/exporters_builder.go:48 msg="Exporter started." component=traces traces_config=default kind=exporter name=otlp/0
ts=2022-06-10T20:49:29Z level=info caller=builder/pipelines_builder.go:223 msg="Pipeline was built." component=traces traces_config=default name=pipeline name=traces
ts=2022-06-10T20:49:29Z level=info caller=builder/pipelines_builder.go:54 msg="Pipeline is starting..." component=traces traces_config=default name=pipeline name=traces
ts=2022-06-10T20:49:29Z level=info caller=builder/pipelines_builder.go:65 msg="Pipeline is started." component=traces traces_config=default name=pipeline name=traces
ts=2022-06-10T20:49:29Z level=info caller=builder/receivers_builder.go:226 msg="Receiver was built." component=traces traces_config=default kind=receiver name=otlp datatype=traces
ts=2022-06-10T20:49:29Z level=info caller=builder/receivers_builder.go:226 msg="Receiver was built." component=traces traces_config=default kind=receiver name=push_receiver datatype=traces
ts=2022-06-10T20:49:29Z level=info caller=builder/receivers_builder.go:68 msg="Receiver is starting..." component=traces traces_config=default kind=receiver name=otlp
ts=2022-06-10T20:49:29Z level=info caller=otlpreceiver/otlp.go:87 msg="Starting HTTP server on endpoint 0.0.0.0:55680" component=traces traces_config=default kind=receiver name=otlp
ts=2022-06-10T20:49:29Z level=info caller=builder/receivers_builder.go:73 msg="Receiver started." component=traces traces_config=default kind=receiver name=otlp
ts=2022-06-10T20:49:29Z level=info caller=builder/receivers_builder.go:68 msg="Receiver is starting..." component=traces traces_config=default kind=receiver name=push_receiver
ts=2022-06-10T20:49:29Z level=info caller=builder/receivers_builder.go:73 msg="Receiver started." component=traces traces_config=default kind=receiver name=push_receiver
ts=2022-06-10T20:49:29.160141007Z caller=manager.go:231 level=debug msg="Applying integrations config changes"
ts=2022-06-10T20:49:29.162172836Z caller=manager.go:228 level=debug msg="Integrations config is unchanged skipping apply"
ts=2022-06-10T20:49:29.162340031Z caller=reporter.go:107 level=info msg="running usage stats reporter"
ts=2022-06-10T20:49:44.153454393Z caller=config_watcher.go:139 level=debug agent=prometheus component=cluster msg="waiting for next reshard interval" last_reshard=2022-06-10T20:49:44.153366114Z next_reshard=2022-06-10T20:50:44.153366114Z remaining=59.999990048s
ts=2022-06-10T20:50:44.153971195Z caller=config_watcher.go:106 level=debug agent=prometheus component=cluster msg="reshard timer ticked, scheduling refresh"
ts=2022-06-10T20:50:44.154072092Z caller=config_watcher.go:147 level=debug agent=prometheus component=cluster msg="successfully scheduled a refresh"
ts=2022-06-10T20:50:44.154086993Z caller=config_watcher.go:139 level=debug agent=prometheus component=cluster msg="waiting for next reshard interval" last_reshard=2022-06-10T20:50:44.154080723Z next_reshard=2022-06-10T20:51:44.154080723Z remaining=59.999998816s
ts=2022-06-10T20:50:44.154103766Z caller=config_watcher.go:163 level=debug agent=prometheus component=cluster msg="refresh skipped because clustering is disabled"
ts=2022-06-10T20:50:44.154113467Z caller=config_watcher.go:139 level=debug agent=prometheus component=cluster msg="waiting for next reshard interval" last_reshard=2022-06-10T20:50:44.154080723Z next_reshard=2022-06-10T20:51:44.154080723Z remaining=59.999968604s
ts=2022-06-10T20:51:44.155162883Z caller=config_watcher.go:106 level=debug agent=prometheus component=cluster msg="reshard timer ticked, scheduling refresh"
ts=2022-06-10T20:51:44.155247031Z caller=config_watcher.go:147 level=debug agent=prometheus component=cluster msg="successfully scheduled a refresh"
ts=2022-06-10T20:51:44.155269105Z caller=config_watcher.go:139 level=debug agent=prometheus component=cluster msg="waiting for next reshard interval" last_reshard=2022-06-10T20:51:44.155258172Z next_reshard=2022-06-10T20:52:44.155258172Z remaining=59.999998769s
ts=2022-06-10T20:51:44.155314634Z caller=config_watcher.go:163 level=debug agent=prometheus component=cluster msg="refresh skipped because clustering is disabled"
ts=2022-06-10T20:51:44.155335545Z caller=config_watcher.go:139 level=debug agent=prometheus component=cluster msg="waiting for next reshard interval" last_reshard=2022-06-10T20:51:44.155258172Z next_reshard=2022-06-10T20:52:44.155258172Z remaining=59.999924474s
Logs from distributor pod:
level=info ts=2022-06-10T22:41:36.55767082Z caller=main.go:191 msg="initialising OpenTracing tracer"
level=info ts=2022-06-10T22:41:36.582889977Z caller=main.go:106 msg="Starting Tempo" version="(version=, branch=HEAD, revision=d3880a979)"
level=info ts=2022-06-10T22:41:36.583853283Z caller=server.go:260 http=[::]:3100 grpc=[::]:9095 msg="server listening on addresses"
ts=2022-06-10T22:41:36Z level=info msg="OTel Shim Logger Initialized" component=tempo
level=info ts=2022-06-10T22:41:36.585850483Z caller=memberlist_client.go:394 msg="Using memberlist cluster node name" name=tempo-tempo-distributed-distributor-54664dc5c4-tw9ph-e26671a8
level=info ts=2022-06-10T22:41:36.587323877Z caller=module_service.go:64 msg=initialising module=server
level=info ts=2022-06-10T22:41:36.587553021Z caller=module_service.go:64 msg=initialising module=memberlist-kv
level=info ts=2022-06-10T22:41:36.587613975Z caller=module_service.go:64 msg=initialising module=overrides
level=info ts=2022-06-10T22:41:36.587762766Z caller=module_service.go:64 msg=initialising module=ring
level=info ts=2022-06-10T22:41:36.589117247Z caller=ring.go:272 msg="ring doesn't exist in KV store yet"
level=info ts=2022-06-10T22:41:36.589265478Z caller=module_service.go:64 msg=initialising module=distributor
ts=2022-06-10T22:41:36Z level=info msg="Starting HTTP server on endpoint 0.0.0.0:55681" component=tempo
level=info ts=2022-06-10T22:41:36.589551498Z caller=app.go:284 msg="Tempo started"
level=info ts=2022-06-10T22:41:36.599493812Z caller=memberlist_client.go:513 msg="joined memberlist cluster" reached_nodes=3
level=warn ts=2022-06-10T22:43:02.854960949Z caller=tcp_transport.go:428 component="memberlist TCPTransport" msg="WriteTo failed" addr=10.215.77.101:7946 err="dial tcp 10.215.77.101:7946: i/o timeout"
ts=2022-06-10T22:43:06.590276921Z caller=memberlist_logger.go:74 level=warn msg="Was able to connect to tempo-tempo-distributed-ingester-0-bd5c0f84 but other probes failed, network may be misconfigured"
level=warn ts=2022-06-10T22:43:07.860665534Z caller=tcp_transport.go:428 component="memberlist TCPTransport" msg="WriteTo failed" addr=10.215.77.101:7946 err="dial tcp 10.215.77.101:7946: i/o timeout"
level=warn ts=2022-06-10T22:43:09.59238309Z caller=tcp_transport.go:428 component="memberlist TCPTransport" msg="WriteTo failed" addr=10.215.77.101:7946 err="dial tcp 10.215.77.101:7946: i/o timeout"
level=warn ts=2022-06-10T22:43:11.5905154Z caller=tcp_transport.go:428 component="memberlist TCPTransport" msg="WriteTo failed" addr=10.215.77.101:7946 err="dial tcp 10.215.77.101:7946: i/o timeout"
ts=2022-06-10T22:43:11.594984883Z caller=memberlist_logger.go:74 level=info msg="Suspect tempo-tempo-distributed-distributor-67f78d648f-gcptg-4ecef0db has failed, no acks received"
level=warn ts=2022-06-10T22:43:12.864858192Z caller=tcp_transport.go:428 component="memberlist TCPTransport" msg="WriteTo failed" addr=10.215.77.101:7946 err="dial tcp 10.215.77.101:7946: i/o timeout"
level=warn ts=2022-06-10T22:43:14.597544152Z caller=tcp_transport.go:428 component="memberlist TCPTransport" msg="WriteTo failed" addr=10.215.77.101:7946 err="dial tcp 10.215.77.101:7946: i/o timeout"
level=warn ts=2022-06-10T22:43:20.174516522Z caller=tcp_transport.go:428 component="memberlist TCPTransport" msg="WriteTo failed" addr=10.215.77.101:7946 err="dial tcp 10.215.77.101:7946: i/o timeout"
ts=2022-06-10T22:43:22.86079914Z caller=memberlist_logger.go:74 level=info msg="Marking tempo-tempo-distributed-distributor-67f78d648f-gcptg-4ecef0db as failed, suspect timeout reached (2 peer confirmations)"
level=warn ts=2022-06-10T22:43:28.590684881Z caller=tcp_transport.go:428 component="memberlist TCPTransport" msg="WriteTo failed" addr=10.215.77.101:7946 err="dial tcp 10.215.77.101:7946: i/o timeout"
ts=2022-06-10T22:43:31.58925098Z caller=memberlist_logger.go:74 level=warn msg="Was able to connect to tempo-tempo-distributed-querier-6c5955ffc5-rpcb6-da45d3d3 but other probes failed, network may be misconfigured"
level=warn ts=2022-06-10T22:43:42.591779235Z caller=tcp_transport.go:428 component="memberlist TCPTransport" msg="WriteTo failed" addr=10.215.77.101:7946 err="dial tcp 10.215.77.101:7946: i/o timeout"
level=warn ts=2022-06-10T22:43:47.592646516Z caller=tcp_transport.go:428 component="memberlist TCPTransport" msg="WriteTo failed" addr=10.215.77.101:7946 err="dial tcp 10.215.77.101:7946: i/o timeout"
level=warn ts=2022-06-10T22:43:53.589591088Z caller=tcp_transport.go:428 component="memberlist TCPTransport" msg="WriteTo failed" addr=10.215.77.101:7946 err="dial tcp 10.215.77.101:7946: i/o timeout"
level=warn ts=2022-06-10T22:44:17.18557509Z caller=tcp_transport.go:428 component="memberlist TCPTransport" msg="WriteTo failed" addr=10.215.77.101:7946 err="dial tcp 10.215.77.101:7946: i/o timeout"
I am not able to figure out what the issue is ? Can you please suggest what I can do to further troubleshoot the issue ?
Thank you.