Files
packager/rust/src/telemetry.rs
2023-08-29 21:34:01 +02:00

158 lines
5.2 KiB
Rust

use std::fmt;
use std::io;
use std::time::Duration;
use axum::Router;
use http::Request;
use tower_http::{classify::ServerErrorsFailureClass, trace::TraceLayer};
use tracing::{Level, Span};
use tracing_subscriber::{
filter::{LevelFilter, Targets},
fmt::{format::Format, Layer},
layer::SubscriberExt,
prelude::*,
registry::Registry,
};
use uuid::Uuid;
use opentelemetry::global;
pub fn otel_init(f: impl FnOnce() -> ()) {
f()
}
pub fn init_tracing() {
// default is the Full format, there is no way to specify this, but it can be
// overridden via builder methods
let stdout_format = Format::default()
.pretty()
.with_ansi(true)
.with_target(true)
.with_level(true)
.with_file(false);
let stdout_layer = Layer::default()
.event_format(stdout_format)
.with_writer(io::stdout);
let stdout_filter = Targets::new()
.with_default(LevelFilter::OFF)
.with_targets(vec![
(env!("CARGO_PKG_NAME"), Level::DEBUG),
// this is for axum requests
("request", Level::DEBUG),
// required for tokio-console as by the docs
// ("tokio", Level::TRACE),
// ("runtime", Level::TRACE),
]);
let stdout_layer = stdout_layer.with_filter(stdout_filter);
let console_layer = console_subscriber::Builder::default().spawn();
global::set_text_map_propagator(opentelemetry_jaeger::Propagator::new());
// Sets up the machinery needed to export data to Jaeger
// There are other OTel crates that provide pipelines for the vendors
// mentioned earlier.
let tracer = opentelemetry_jaeger::new_agent_pipeline()
.with_service_name(env!("CARGO_PKG_NAME"))
.install_simple()
.unwrap();
let opentelemetry_filter = Targets::new()
.with_default(LevelFilter::OFF)
.with_targets(vec![
(env!("CARGO_PKG_NAME"), Level::DEBUG),
// this is for axum requests
("request", Level::DEBUG),
// required for tokio-console as by the docs
// ("tokio", Level::TRACE),
// ("runtime", Level::TRACE),
]);
let opentelemetry = tracing_opentelemetry::layer()
.with_tracer(tracer)
.with_filter(opentelemetry_filter);
let registry = Registry::default()
.with(console_layer)
.with(opentelemetry)
// just an example, you can actuall pass Options here for layers that might be
// set/unset at runtime
.with(Some(stdout_layer))
.with(None::<Layer<_>>);
tracing::subscriber::set_global_default(registry).unwrap();
}
struct Latency(Duration);
impl fmt::Display for Latency {
fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
write!(f, "{}", self.0.as_micros())
}
}
pub fn init_request_tracing(router: Router) -> Router {
router.layer(
TraceLayer::new_for_http()
.make_span_with(|_request: &Request<_>| {
let request_id = Uuid::new_v4();
tracing::debug_span!(
target: "request",
"request",
%request_id,
)
})
.on_request(|request: &Request<_>, _span: &Span| {
let request_headers = request.headers();
let http_version = request.version();
tracing::debug!(
target: "request",
method = request.method().as_str(),
path = request.uri().path(),
?http_version,
?request_headers,
"request received",
);
})
.on_response(
|response: &axum::response::Response, latency: Duration, _span: &Span| {
let response_headers = response.headers();
let latency = Latency(latency);
tracing::debug!(
target: "request",
%latency,
status = response.status().as_str(),
?response_headers,
"finished processing request",
);
},
)
.on_failure(
|error: ServerErrorsFailureClass, latency: Duration, _span: &Span| {
let latency = Latency(latency);
match error {
ServerErrorsFailureClass::StatusCode(code) => {
tracing::error!(
target: "request",
%latency,
"request failed with error response {}",
code,
);
}
ServerErrorsFailureClass::Error(message) => {
tracing::error!(
target: "request",
%latency,
"request failed: {}",
message,
);
}
}
},
),
)
}