From 8d0224a9c23a05dc136ea6b2fc12cff69f64b633 Mon Sep 17 00:00:00 2001 From: Filipe Azevedo Date: Fri, 14 Feb 2025 13:59:38 +0000 Subject: [PATCH] query tracing --- Cargo.lock | 20 +- Cargo.toml | 15 +- graph/build.rs | 6 + graph/proto/tracing.proto | 17 ++ graph/src/components/mod.rs | 2 + graph/src/components/tracing.rs | 123 ++++++++++ graph/src/data/query/trace.rs | 4 +- graph/src/grpc/mod.rs | 4 + graph/src/grpc/pb/graph.tracing.v1.rs | 312 ++++++++++++++++++++++++++ graph/src/grpc/pb/mod.rs | 8 + graph/src/lib.rs | 2 + node/Cargo.toml | 1 + node/src/main.rs | 20 +- server/grpc/Cargo.toml | 15 ++ server/grpc/src/lib.rs | 128 +++++++++++ store/postgres/src/lib.rs | 8 + store/postgres/src/relational.rs | 38 +++- 17 files changed, 709 insertions(+), 14 deletions(-) create mode 100644 graph/proto/tracing.proto create mode 100644 graph/src/components/tracing.rs create mode 100644 graph/src/grpc/mod.rs create mode 100644 graph/src/grpc/pb/graph.tracing.v1.rs create mode 100644 graph/src/grpc/pb/mod.rs create mode 100644 server/grpc/Cargo.toml create mode 100644 server/grpc/src/lib.rs diff --git a/Cargo.lock b/Cargo.lock index b2ae24e0169..fd96751be75 100644 --- a/Cargo.lock +++ b/Cargo.lock @@ -643,9 +643,9 @@ checksum = "1fd0f2584146f6f2ef48085050886acf353beff7305ebd1ae69500e27c67f64b" [[package]] name = "bytes" -version = "1.6.0" +version = "1.10.0" source = "registry+https://github.com/rust-lang/crates.io-index" -checksum = "514de17de45fdb8dc022b1a7975556c53c86f9f0aa5f534b98977b171857c2c9" +checksum = "f61dac84819c6588b558454b194026eb1f09c293b9036ae9b159e74e73ab6cf9" dependencies = [ "serde", ] @@ -2017,6 +2017,7 @@ dependencies = [ "graph-chain-substreams", "graph-core", "graph-graphql", + "graph-server-grpc", "graph-server-http", "graph-server-index-node", "graph-server-json-rpc", @@ -2079,6 +2080,16 @@ dependencies = [ "wasmtime", ] +[[package]] +name = "graph-server-grpc" +version = "0.36.0" +dependencies = [ + "graph", + "graph-store-postgres", + "tokio", + "tonic", +] + [[package]] name = "graph-server-http" version = "0.36.0" @@ -2606,9 +2617,9 @@ dependencies = [ [[package]] name = "hyper-util" -version = "0.1.6" +version = "0.1.10" source = "registry+https://github.com/rust-lang/crates.io-index" -checksum = "3ab92f4f49ee4fb4f997c784b7a2e0fa70050211e0b6a287f898c3c9785ca956" +checksum = "df2dcfbe0677734ab2f3ffa7fa7bfd4706bfdc1ef393f2ee30184aed67e631b4" dependencies = [ "bytes", "futures-channel", @@ -2619,7 +2630,6 @@ dependencies = [ "pin-project-lite", "socket2", "tokio", - "tower 0.4.13 (registry+https://github.com/rust-lang/crates.io-index)", "tower-service 0.3.2 (registry+https://github.com/rust-lang/crates.io-index)", "tracing", ] diff --git a/Cargo.toml b/Cargo.toml index 751a19d6213..7dd13caf6c1 100644 --- a/Cargo.toml +++ b/Cargo.toml @@ -34,7 +34,15 @@ bs58 = "0.5.1" chrono = "0.4.38" clap = { version = "4.5.4", features = ["derive", "env"] } derivative = "2.2.0" -diesel = { version = "2.2.4", features = ["postgres", "serde_json", "numeric", "r2d2", "chrono", "uuid", "i-implement-a-third-party-backend-and-opt-into-breaking-changes"] } +diesel = { version = "2.2.4", features = [ + "postgres", + "serde_json", + "numeric", + "r2d2", + "chrono", + "uuid", + "i-implement-a-third-party-backend-and-opt-into-breaking-changes", +] } diesel-derive-enum = { version = "2.1.0", features = ["postgres"] } diesel-dynamic-schema = { version = "0.2.1", features = ["postgres"] } diesel_derives = "2.1.4" @@ -56,7 +64,10 @@ serde_derive = "1.0.125" serde_json = { version = "1.0", features = ["arbitrary_precision"] } serde_regex = "1.1.0" serde_yaml = "0.9.21" -slog = { version = "2.7.0", features = ["release_max_level_trace", "max_level_trace"] } +slog = { version = "2.7.0", features = [ + "release_max_level_trace", + "max_level_trace", +] } sqlparser = "0.46.0" strum = { version = "0.26", features = ["derive"] } syn = { version = "2.0.66", features = ["full"] } diff --git a/graph/build.rs b/graph/build.rs index 3cc00c0dc07..79f63c5eb58 100644 --- a/graph/build.rs +++ b/graph/build.rs @@ -25,4 +25,10 @@ fn main() { .out_dir("src/substreams_rpc") .compile(&["proto/substreams-rpc.proto"], &["proto"]) .expect("Failed to compile Substreams RPC proto(s)"); + + tonic_build::configure() + .out_dir("src/grpc/pb") + .include_file("mod.rs") + .compile(&["proto/tracing.proto"], &["proto"]) + .expect("Failed to compile Tracing proto(s)"); } diff --git a/graph/proto/tracing.proto b/graph/proto/tracing.proto new file mode 100644 index 00000000000..9cf036d126e --- /dev/null +++ b/graph/proto/tracing.proto @@ -0,0 +1,17 @@ +syntax = "proto3"; + +package graph.tracing.v1; + +service Stream { + rpc QueryTrace(Request) returns (stream Trace); +} + +message Request { + int32 deployment_id = 1; +} + +message Trace { + int32 deployment_id = 1; + string query = 2; + uint64 duration_millis = 3; +} diff --git a/graph/src/components/mod.rs b/graph/src/components/mod.rs index 8abdc96f0b0..5561b3142ec 100644 --- a/graph/src/components/mod.rs +++ b/graph/src/components/mod.rs @@ -60,6 +60,8 @@ pub mod metrics; /// Components dealing with versioning pub mod versions; +pub mod tracing; + /// A component that receives events of type `T`. pub trait EventConsumer { /// Get the event sink. diff --git a/graph/src/components/tracing.rs b/graph/src/components/tracing.rs new file mode 100644 index 00000000000..a4d925b6e0f --- /dev/null +++ b/graph/src/components/tracing.rs @@ -0,0 +1,123 @@ +use std::{collections::HashMap, sync::atomic::AtomicBool}; + +use tokio::sync::{mpsc, RwLock}; + +use super::store::DeploymentId; + +const DEFAULT_BUFFER_SIZE: usize = 100; + +#[derive(Debug)] +pub struct Subscriptions { + inner: RwLock>>, +} + +/// A control structure for managing tracing subscriptions. +#[derive(Debug)] +pub struct TracingControl { + enabled: AtomicBool, + subscriptions: Subscriptions, + default_buffer_size: usize, +} + +impl Default for TracingControl { + fn default() -> Self { + Self { + enabled: AtomicBool::new(false), + subscriptions: Subscriptions { + inner: RwLock::new(HashMap::new()), + }, + default_buffer_size: DEFAULT_BUFFER_SIZE, + } + } +} + +impl TracingControl { + pub fn new(default_buffer_size: Option) -> Self { + Self { + enabled: AtomicBool::new(false), + subscriptions: Subscriptions { + inner: RwLock::new(HashMap::new()), + }, + default_buffer_size: default_buffer_size.unwrap_or(DEFAULT_BUFFER_SIZE), + } + } + + /// Creates a channel sender for the given deployment ID. Only one subscription + /// can exist for a given deployment ID. If tracing is disabled or no subscription + /// exists, it will return None. Calling producer when a dead subscription exists + /// will incur a cleanup cost. + pub async fn producer(&self, key: DeploymentId) -> Option> { + if !self.enabled.load(std::sync::atomic::Ordering::Relaxed) { + return None; + } + + let subs = self.subscriptions.inner.read().await; + let tx = subs.get(&key); + + match tx { + Some(tx) if tx.is_closed() => { + drop(subs); + let mut subs = self.subscriptions.inner.write().await; + subs.remove(&key); + + if subs.is_empty() { + self.enabled + .store(false, std::sync::atomic::Ordering::Relaxed); + } + + None + } + None => None, + tx => tx.cloned(), + } + } + pub async fn subscribe_with_chan_size( + &self, + key: DeploymentId, + buffer_size: usize, + ) -> mpsc::Receiver { + let (tx, rx) = mpsc::channel(buffer_size); + let mut guard = self.subscriptions.inner.write().await; + guard.insert(key, tx); + self.enabled + .store(true, std::sync::atomic::Ordering::Relaxed); + + rx + } + + /// Creates a new subscription for a given deployment ID. If a subscription already + /// exists, it will be replaced. + pub async fn subscribe(&self, key: DeploymentId) -> mpsc::Receiver { + self.subscribe_with_chan_size(key, self.default_buffer_size) + .await + } +} + +#[cfg(test)] +mod test { + + use super::*; + use std::sync::atomic::Ordering::Relaxed; + use std::sync::Arc; + + #[tokio::test] + async fn test_tracing_control() { + let control: TracingControl<()> = TracingControl::default(); + let control = Arc::new(control); + assert_eq!(false, control.enabled.load(Relaxed)); + + let tx = control.producer(DeploymentId(123)).await; + assert!(tx.is_none()); + + let rx = control.subscribe(DeploymentId(123)).await; + assert_eq!(true, control.enabled.load(Relaxed)); + + drop(rx); + let tx = control.producer(DeploymentId(123)).await; + assert!(tx.is_none()); + assert_eq!(false, control.enabled.load(Relaxed)); + + _ = control.subscribe(DeploymentId(123)).await; + assert_eq!(true, control.enabled.load(Relaxed)); + } +} diff --git a/graph/src/data/query/trace.rs b/graph/src/data/query/trace.rs index cf2d153dca4..dc8264a10df 100644 --- a/graph/src/data/query/trace.rs +++ b/graph/src/data/query/trace.rs @@ -14,7 +14,7 @@ lazy_static! { pub static ref TRACE_NONE: Arc = Arc::new(Trace::None); } -#[derive(Debug, CacheWeight)] +#[derive(Debug, CacheWeight, Clone)] pub struct TraceWithCacheStatus { pub trace: Arc, pub cache_status: CacheStatus, @@ -35,7 +35,7 @@ impl HttpTrace { } } -#[derive(Debug, CacheWeight)] +#[derive(Debug, CacheWeight, Clone)] pub enum Trace { None, Root { diff --git a/graph/src/grpc/mod.rs b/graph/src/grpc/mod.rs new file mode 100644 index 00000000000..6189c18f361 --- /dev/null +++ b/graph/src/grpc/mod.rs @@ -0,0 +1,4 @@ +use pb::graph::tracing::v1::{Request, Trace}; +use tonic::async_trait; + +pub mod pb; diff --git a/graph/src/grpc/pb/graph.tracing.v1.rs b/graph/src/grpc/pb/graph.tracing.v1.rs new file mode 100644 index 00000000000..472c0d14a26 --- /dev/null +++ b/graph/src/grpc/pb/graph.tracing.v1.rs @@ -0,0 +1,312 @@ +// This file is @generated by prost-build. +#[allow(clippy::derive_partial_eq_without_eq)] +#[derive(Clone, PartialEq, ::prost::Message)] +pub struct Request { + #[prost(int32, tag = "1")] + pub deployment_id: i32, +} +#[allow(clippy::derive_partial_eq_without_eq)] +#[derive(Clone, PartialEq, ::prost::Message)] +pub struct Trace { + #[prost(int32, tag = "1")] + pub deployment_id: i32, + #[prost(string, tag = "2")] + pub query: ::prost::alloc::string::String, + #[prost(uint64, tag = "3")] + pub duration_millis: u64, +} +/// Generated client implementations. +pub mod stream_client { + #![allow(unused_variables, dead_code, missing_docs, clippy::let_unit_value)] + use tonic::codegen::*; + use tonic::codegen::http::Uri; + #[derive(Debug, Clone)] + pub struct StreamClient { + inner: tonic::client::Grpc, + } + impl StreamClient { + /// Attempt to create a new client by connecting to a given endpoint. + pub async fn connect(dst: D) -> Result + where + D: TryInto, + D::Error: Into, + { + let conn = tonic::transport::Endpoint::new(dst)?.connect().await?; + Ok(Self::new(conn)) + } + } + impl StreamClient + where + T: tonic::client::GrpcService, + T::Error: Into, + T::ResponseBody: Body + Send + 'static, + ::Error: Into + Send, + { + pub fn new(inner: T) -> Self { + let inner = tonic::client::Grpc::new(inner); + Self { inner } + } + pub fn with_origin(inner: T, origin: Uri) -> Self { + let inner = tonic::client::Grpc::with_origin(inner, origin); + Self { inner } + } + pub fn with_interceptor( + inner: T, + interceptor: F, + ) -> StreamClient> + where + F: tonic::service::Interceptor, + T::ResponseBody: Default, + T: tonic::codegen::Service< + http::Request, + Response = http::Response< + >::ResponseBody, + >, + >, + , + >>::Error: Into + Send + Sync, + { + StreamClient::new(InterceptedService::new(inner, interceptor)) + } + /// Compress requests with the given encoding. + /// + /// This requires the server to support it otherwise it might respond with an + /// error. + #[must_use] + pub fn send_compressed(mut self, encoding: CompressionEncoding) -> Self { + self.inner = self.inner.send_compressed(encoding); + self + } + /// Enable decompressing responses. + #[must_use] + pub fn accept_compressed(mut self, encoding: CompressionEncoding) -> Self { + self.inner = self.inner.accept_compressed(encoding); + self + } + /// Limits the maximum size of a decoded message. + /// + /// Default: `4MB` + #[must_use] + pub fn max_decoding_message_size(mut self, limit: usize) -> Self { + self.inner = self.inner.max_decoding_message_size(limit); + self + } + /// Limits the maximum size of an encoded message. + /// + /// Default: `usize::MAX` + #[must_use] + pub fn max_encoding_message_size(mut self, limit: usize) -> Self { + self.inner = self.inner.max_encoding_message_size(limit); + self + } + pub async fn query_trace( + &mut self, + request: impl tonic::IntoRequest, + ) -> std::result::Result< + tonic::Response>, + tonic::Status, + > { + self.inner + .ready() + .await + .map_err(|e| { + tonic::Status::new( + tonic::Code::Unknown, + format!("Service was not ready: {}", e.into()), + ) + })?; + let codec = tonic::codec::ProstCodec::default(); + let path = http::uri::PathAndQuery::from_static( + "/graph.tracing.v1.Stream/QueryTrace", + ); + let mut req = request.into_request(); + req.extensions_mut() + .insert(GrpcMethod::new("graph.tracing.v1.Stream", "QueryTrace")); + self.inner.server_streaming(req, path, codec).await + } + } +} +/// Generated server implementations. +pub mod stream_server { + #![allow(unused_variables, dead_code, missing_docs, clippy::let_unit_value)] + use tonic::codegen::*; + /// Generated trait containing gRPC methods that should be implemented for use with StreamServer. + #[async_trait] + pub trait Stream: Send + Sync + 'static { + /// Server streaming response type for the QueryTrace method. + type QueryTraceStream: tonic::codegen::tokio_stream::Stream< + Item = std::result::Result, + > + + Send + + 'static; + async fn query_trace( + &self, + request: tonic::Request, + ) -> std::result::Result, tonic::Status>; + } + #[derive(Debug)] + pub struct StreamServer { + inner: _Inner, + accept_compression_encodings: EnabledCompressionEncodings, + send_compression_encodings: EnabledCompressionEncodings, + max_decoding_message_size: Option, + max_encoding_message_size: Option, + } + struct _Inner(Arc); + impl StreamServer { + pub fn new(inner: T) -> Self { + Self::from_arc(Arc::new(inner)) + } + pub fn from_arc(inner: Arc) -> Self { + let inner = _Inner(inner); + Self { + inner, + accept_compression_encodings: Default::default(), + send_compression_encodings: Default::default(), + max_decoding_message_size: None, + max_encoding_message_size: None, + } + } + pub fn with_interceptor( + inner: T, + interceptor: F, + ) -> InterceptedService + where + F: tonic::service::Interceptor, + { + InterceptedService::new(Self::new(inner), interceptor) + } + /// Enable decompressing requests with the given encoding. + #[must_use] + pub fn accept_compressed(mut self, encoding: CompressionEncoding) -> Self { + self.accept_compression_encodings.enable(encoding); + self + } + /// Compress responses with the given encoding, if the client supports it. + #[must_use] + pub fn send_compressed(mut self, encoding: CompressionEncoding) -> Self { + self.send_compression_encodings.enable(encoding); + self + } + /// Limits the maximum size of a decoded message. + /// + /// Default: `4MB` + #[must_use] + pub fn max_decoding_message_size(mut self, limit: usize) -> Self { + self.max_decoding_message_size = Some(limit); + self + } + /// Limits the maximum size of an encoded message. + /// + /// Default: `usize::MAX` + #[must_use] + pub fn max_encoding_message_size(mut self, limit: usize) -> Self { + self.max_encoding_message_size = Some(limit); + self + } + } + impl tonic::codegen::Service> for StreamServer + where + T: Stream, + B: Body + Send + 'static, + B::Error: Into + Send + 'static, + { + type Response = http::Response; + type Error = std::convert::Infallible; + type Future = BoxFuture; + fn poll_ready( + &mut self, + _cx: &mut Context<'_>, + ) -> Poll> { + Poll::Ready(Ok(())) + } + fn call(&mut self, req: http::Request) -> Self::Future { + let inner = self.inner.clone(); + match req.uri().path() { + "/graph.tracing.v1.Stream/QueryTrace" => { + #[allow(non_camel_case_types)] + struct QueryTraceSvc(pub Arc); + impl tonic::server::ServerStreamingService + for QueryTraceSvc { + type Response = super::Trace; + type ResponseStream = T::QueryTraceStream; + type Future = BoxFuture< + tonic::Response, + tonic::Status, + >; + fn call( + &mut self, + request: tonic::Request, + ) -> Self::Future { + let inner = Arc::clone(&self.0); + let fut = async move { + ::query_trace(&inner, request).await + }; + Box::pin(fut) + } + } + let accept_compression_encodings = self.accept_compression_encodings; + let send_compression_encodings = self.send_compression_encodings; + let max_decoding_message_size = self.max_decoding_message_size; + let max_encoding_message_size = self.max_encoding_message_size; + let inner = self.inner.clone(); + let fut = async move { + let inner = inner.0; + let method = QueryTraceSvc(inner); + let codec = tonic::codec::ProstCodec::default(); + let mut grpc = tonic::server::Grpc::new(codec) + .apply_compression_config( + accept_compression_encodings, + send_compression_encodings, + ) + .apply_max_message_size_config( + max_decoding_message_size, + max_encoding_message_size, + ); + let res = grpc.server_streaming(method, req).await; + Ok(res) + }; + Box::pin(fut) + } + _ => { + Box::pin(async move { + Ok( + http::Response::builder() + .status(200) + .header("grpc-status", "12") + .header("content-type", "application/grpc") + .body(empty_body()) + .unwrap(), + ) + }) + } + } + } + } + impl Clone for StreamServer { + fn clone(&self) -> Self { + let inner = self.inner.clone(); + Self { + inner, + accept_compression_encodings: self.accept_compression_encodings, + send_compression_encodings: self.send_compression_encodings, + max_decoding_message_size: self.max_decoding_message_size, + max_encoding_message_size: self.max_encoding_message_size, + } + } + } + impl Clone for _Inner { + fn clone(&self) -> Self { + Self(Arc::clone(&self.0)) + } + } + impl std::fmt::Debug for _Inner { + fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result { + write!(f, "{:?}", self.0) + } + } + impl tonic::server::NamedService for StreamServer { + const NAME: &'static str = "graph.tracing.v1.Stream"; + } +} diff --git a/graph/src/grpc/pb/mod.rs b/graph/src/grpc/pb/mod.rs new file mode 100644 index 00000000000..611e514e404 --- /dev/null +++ b/graph/src/grpc/pb/mod.rs @@ -0,0 +1,8 @@ +// This file is @generated by prost-build. +pub mod graph { + pub mod tracing { + pub mod v1 { + include!("graph.tracing.v1.rs"); + } + } +} diff --git a/graph/src/lib.rs b/graph/src/lib.rs index 04872aab196..3de1fea3d9d 100644 --- a/graph/src/lib.rs +++ b/graph/src/lib.rs @@ -37,6 +37,8 @@ pub mod env; pub mod ipfs; +pub mod grpc; + /// Wrapper for spawning tasks that abort on panic, which is our default. mod task_spawn; pub use task_spawn::{ diff --git a/node/Cargo.toml b/node/Cargo.toml index 820ed8405a8..0c9b641f1da 100644 --- a/node/Cargo.toml +++ b/node/Cargo.toml @@ -32,6 +32,7 @@ graph-server-http = { path = "../server/http" } graph-server-index-node = { path = "../server/index-node" } graph-server-json-rpc = { path = "../server/json-rpc" } graph-server-websocket = { path = "../server/websocket" } +graph-server-grpc = { path = "../server/grpc" } graph-server-metrics = { path = "../server/metrics" } graph-store-postgres = { path = "../store/postgres" } graphman-server = { workspace = true } diff --git a/node/src/main.rs b/node/src/main.rs index 870cce97318..9c3f00d34ae 100644 --- a/node/src/main.rs +++ b/node/src/main.rs @@ -1,5 +1,6 @@ use clap::Parser as _; use git_testament::{git_testament, render_testament}; +use graph::components::store::DeploymentId; use graph::futures01::Future as _; use graph::futures03::compat::Future01CompatExt; use graph::futures03::future::TryFutureExt; @@ -30,8 +31,8 @@ use graph_server_json_rpc::JsonRpcServer; use graph_server_metrics::PrometheusMetricsServer; use graph_server_websocket::SubscriptionServer as GraphQLSubscriptionServer; use graph_store_postgres::connection_pool::ConnectionPool; -use graph_store_postgres::Store; use graph_store_postgres::{register_jobs as register_store_jobs, NotificationSender}; +use graph_store_postgres::{Store, TRACING_CONTROL}; use graphman_server::GraphmanServer; use graphman_server::GraphmanServerConfig; use std::io::{BufRead, BufReader}; @@ -516,6 +517,10 @@ async fn main() { // Run the index node server graph::spawn(async move { index_node_server.start(index_node_port).await }); + graph::spawn(async move { + graph_server_grpc::start(8888).await.unwrap(); + }); + graph::spawn(async move { metrics_server .start(metrics_port) @@ -559,6 +564,19 @@ async fn main() { } }); + let mut rx = TRACING_CONTROL.subscribe(DeploymentId(1)).await; + loop { + let trace = rx.recv().await; + match trace { + Some(trace) => { + info!(&logger, "#### trace: {:?}", trace); + } + None => { + break; + } + } + } + graph::futures03::future::pending::<()>().await; } diff --git a/server/grpc/Cargo.toml b/server/grpc/Cargo.toml new file mode 100644 index 00000000000..10e101ffcd9 --- /dev/null +++ b/server/grpc/Cargo.toml @@ -0,0 +1,15 @@ +[package] +name = "graph-server-grpc" +version.workspace = true +edition.workspace = true +authors.workspace = true +readme.workspace = true +homepage.workspace = true +repository.workspace = true +license.workspace = true + +[dependencies] +graph = { path = "../../graph" } +graph-store-postgres = { path = "../../store/postgres" } +tonic.workspace = true +tokio.workspace = true diff --git a/server/grpc/src/lib.rs b/server/grpc/src/lib.rs new file mode 100644 index 00000000000..85773a065d0 --- /dev/null +++ b/server/grpc/src/lib.rs @@ -0,0 +1,128 @@ +pub struct TracingServer; + +use std::pin::Pin; + +use graph::{ + futures03::Stream, + grpc::pb::graph::tracing::v1::{ + stream_server::{Stream as StreamProto, StreamServer}, + Request, Trace as TraceProto, + }, + tokio_stream::wrappers::ReceiverStream, +}; +use graph_store_postgres::TRACING_CONTROL; +use tonic::{async_trait, Status}; + +type ResponseStream = Pin> + Send>>; + +#[async_trait] +impl StreamProto for TracingServer { + type QueryTraceStream = ResponseStream; + + async fn query_trace( + &self, + request: tonic::Request, + ) -> std::result::Result, tonic::Status> { + let Request { deployment_id } = request.into_inner(); + + let mut rx = TRACING_CONTROL + .subscribe(graph::components::store::DeploymentId(deployment_id)) + .await; + + // let stream: Pin>> = unfold(rx, |mut rx| async move { + // rx.recv().await.map(|trace| { + // let trace = match trace { + // graph::data::query::Trace::None => vec![], + // graph::data::query::Trace::Root { + // query, + // variables, + // query_id, + // setup, + // elapsed, + // query_parsing, + // blocks, + // } => vec![], + // graph::data::query::Trace::Block { + // block, + // elapsed, + // permit_wait, + // children, + // } => vec![], + // graph::data::query::Trace::Query { + // query, + // elapsed, + // conn_wait, + // permit_wait, + // entity_count, + // children, + // } => vec![TraceProto { + // deployment_id, + // query, + // duration_millis: elapsed.as_millis() as u64, + // }], + // }; + + // (trace, rx) + // }) + // }) + // .boxed(); + // .flatten(); + // + let (tx, rx2) = tokio::sync::mpsc::channel(100); + + tokio::spawn(async move { + while let Some(result) = rx.recv().await { + let out = match result { + graph::data::query::Trace::None => continue, + graph::data::query::Trace::Root { + query: _, + variables: _, + query_id: _, + setup: _, + elapsed: _, + query_parsing: _, + blocks: _, + } => continue, + graph::data::query::Trace::Block { + block: _, + elapsed: _, + permit_wait: _, + children: _, + } => continue, + graph::data::query::Trace::Query { + query, + elapsed, + conn_wait: _, + permit_wait: _, + entity_count: _, + children: _, + } => TraceProto { + deployment_id, + query, + duration_millis: elapsed.as_millis() as u64, + }, + }; + + tx.send(Ok(out)).await.unwrap(); + } + println!("\tstream ended"); + }); + + let out_stream = ReceiverStream::new(rx2); + + Ok(tonic::Response::new( + Box::pin(out_stream) as Self::QueryTraceStream + )) + } +} + +pub async fn start(port: u16) -> Result<(), Box> { + let addr = format!("[::]:{}", port).parse()?; + println!("gRPC server listening on {}", addr); + tonic::transport::Server::builder() + .add_service(StreamServer::new(TracingServer)) + .serve(addr) + .await?; + + Ok(()) +} diff --git a/store/postgres/src/lib.rs b/store/postgres/src/lib.rs index 759e8601313..f9888f39124 100644 --- a/store/postgres/src/lib.rs +++ b/store/postgres/src/lib.rs @@ -11,6 +11,14 @@ extern crate diesel_migrations; #[macro_use] extern crate diesel_derive_enum; +use graph::components::tracing::TracingControl; +use graph::data::query::Trace; +use graph::prelude::lazy_static; + +lazy_static! { + pub static ref TRACING_CONTROL: TracingControl = TracingControl::default(); +} + mod advisory_lock; mod block_range; mod block_store; diff --git a/store/postgres/src/relational.rs b/store/postgres/src/relational.rs index de7e6895083..99c747186f4 100644 --- a/store/postgres/src/relational.rs +++ b/store/postgres/src/relational.rs @@ -33,7 +33,6 @@ use graph::blockchain::BlockTime; use graph::cheap_clone::CheapClone; use graph::components::store::write::{RowGroup, WriteChunk}; use graph::components::subgraph::PoICausalityRegion; -use graph::constraint_violation; use graph::data::graphql::TypeExt as _; use graph::data::query::Trace; use graph::data::value::Word; @@ -43,6 +42,7 @@ use graph::schema::{ EntityKey, EntityType, Field, FulltextConfig, FulltextDefinition, InputSchema, }; use graph::slog::warn; +use graph::{constraint_violation, tokio}; use index::IndexList; use inflector::Inflector; use itertools::Itertools; @@ -54,6 +54,7 @@ use std::fmt::{self, Write}; use std::ops::Range; use std::str::FromStr; use std::sync::{Arc, Mutex}; +use std::thread; use std::time::{Duration, Instant}; use crate::relational::value::{FromOidRow, OidRow}; @@ -79,7 +80,7 @@ use graph::prelude::{ use crate::block_range::{BoundSide, BLOCK_COLUMN, BLOCK_RANGE_COLUMN}; pub use crate::catalog::Catalog; use crate::connection_pool::ForeignServer; -use crate::{catalog, deployment}; +use crate::{catalog, deployment, TRACING_CONTROL}; use self::rollup::Rollup; @@ -824,11 +825,12 @@ impl Layout { elapsed: Duration, entity_count: usize, trace: bool, + trace_sender: Option>, ) -> Trace { // 20kB const MAXLEN: usize = 20_480; - if !ENV_VARS.log_sql_timing() && !trace { + if !ENV_VARS.log_sql_timing() && !trace && trace_sender.is_none() { return Trace::None; } @@ -840,6 +842,16 @@ impl Layout { Trace::None }; + match (&trace, trace_sender) { + (Trace::None, Some(sender)) => { + let _ = sender.try_send(Trace::query(&text, elapsed, entity_count)); + } + (trace, Some(sender)) => { + let _ = sender.try_send(trace.clone()); + } + (_, None) => {} + } + if ENV_VARS.log_sql_timing() { // If the query + bind variables is more than MAXLEN, truncate it; // this will happen when queries have very large bind variables @@ -910,7 +922,25 @@ impl Layout { )), } })?; - let trace = log_query_timing(logger, &query_clone, start.elapsed(), values.len(), trace); + + let handle = tokio::runtime::Handle::current(); + let id = self.site.id.into(); + + // TODO: Find a better way to do this. + let trace_sender = thread::spawn(move || { + handle.block_on(async move { TRACING_CONTROL.producer(id).await }) + }) + .join() + .unwrap(); + + let trace = log_query_timing( + logger, + &query_clone, + start.elapsed(), + values.len(), + trace, + trace_sender, + ); let parent_type = filter_collection.parent_type()?.map(ColumnType::from); values