Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

chore(gcp_stackdriver_logs sink): refactor to new style #18335

Merged
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
Show all changes
47 commits
Select commit Hold shift + click to select a range
a16de09
First iteration, unit tests pass.
neuronull Aug 9, 2023
969653d
Extract common Service functionality for HTTP based stream sinks.
neuronull Aug 9, 2023
5dc4921
docs touchup
neuronull Aug 9, 2023
10fef82
Touch ups
neuronull Aug 9, 2023
d320b10
spell checker
neuronull Aug 9, 2023
a496821
fix rust doc
neuronull Aug 10, 2023
44c99b3
hopefully fix encoding regression
neuronull Aug 10, 2023
4aa3eca
Try item sized batching
neuronull Aug 14, 2023
68ff83c
Refactor a bit
neuronull Aug 15, 2023
98c3113
cleanup
neuronull Aug 15, 2023
152de8a
cleanup
neuronull Aug 15, 2023
1e696d4
doc clean up
neuronull Aug 16, 2023
06381de
extract common service code for re use in other HTTP sinks
neuronull Aug 16, 2023
24a28f7
Merge branch 'master' into neuronull/sink_newstyle_refactor_http
neuronull Aug 16, 2023
9efe546
feeback sw
neuronull Aug 17, 2023
c22f4d3
Merge branch 'master' into neuronull/sink_newstyle_refactor_http
neuronull Aug 17, 2023
2ea6181
duplicate
neuronull Aug 17, 2023
2bfe0d0
feedback ds
neuronull Aug 18, 2023
b2fac79
feedback ds
neuronull Aug 18, 2023
5316602
clippy
neuronull Aug 18, 2023
38969d2
unit tests pass
neuronull Aug 21, 2023
2fe1f05
First iteration, unit tests pass.
neuronull Aug 9, 2023
fe2e4a4
Extract common Service functionality for HTTP based stream sinks.
neuronull Aug 9, 2023
3fc23a1
docs touchup
neuronull Aug 9, 2023
fcc1851
Touch ups
neuronull Aug 9, 2023
a521b2f
spell checker
neuronull Aug 9, 2023
b99b9a7
fix rust doc
neuronull Aug 10, 2023
d9c6acc
hopefully fix encoding regression
neuronull Aug 10, 2023
ebf8e77
Try item sized batching
neuronull Aug 14, 2023
8f8d2a2
Refactor a bit
neuronull Aug 15, 2023
8f91d17
cleanup
neuronull Aug 15, 2023
527a919
cleanup
neuronull Aug 15, 2023
89d3909
doc clean up
neuronull Aug 16, 2023
dfe2335
extract common service code for re use in other HTTP sinks
neuronull Aug 16, 2023
43ae60f
feeback sw
neuronull Aug 17, 2023
7c3e56d
duplicate
neuronull Aug 17, 2023
d4337f8
feedback ds
neuronull Aug 18, 2023
a66a924
feedback ds
neuronull Aug 18, 2023
bc6ec0d
clippy
neuronull Aug 18, 2023
b58e48f
unit tests pass
neuronull Aug 21, 2023
17aacf4
Merge branch 'neuronull/sink_newstyle_refactor_gcp_stackdriver_logs' …
neuronull Aug 21, 2023
e6ddb44
cleanup
neuronull Aug 21, 2023
517c692
use the common batcher
neuronull Aug 21, 2023
20c9199
Merge branch 'master' into neuronull/sink_newstyle_refactor_gcp_stack…
neuronull Aug 21, 2023
4933945
merge conflicts
neuronull Aug 21, 2023
6571e8d
clippy
neuronull Aug 21, 2023
71c5c03
feedback ds
neuronull Aug 23, 2023
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
2 changes: 1 addition & 1 deletion src/sinks/gcp/mod.rs
Original file line number Diff line number Diff line change
Expand Up @@ -6,7 +6,7 @@ use vector_config::configurable_component;
pub mod chronicle_unstructured;
pub mod cloud_storage;
pub mod pubsub;
pub mod stackdriver_logs;
pub mod stackdriver;
pub mod stackdriver_metrics;

/// A monitored resource.
Expand Down
278 changes: 278 additions & 0 deletions src/sinks/gcp/stackdriver/logs/config.rs
Original file line number Diff line number Diff line change
@@ -0,0 +1,278 @@
//! Configuration for the `gcp_stackdriver_logs` sink.

use crate::{
gcp::{GcpAuthConfig, GcpAuthenticator, Scope},
http::HttpClient,
schema,
sinks::{
gcs_common::config::healthcheck_response,
prelude::*,
util::{
http::{http_response_retry_logic, HttpService},
BoxedRawValue, RealtimeSizeBasedDefaultBatchSettings,
},
},
};
use http::{Request, Uri};
use hyper::Body;
use lookup::lookup_v2::ConfigValuePath;
use snafu::Snafu;
use std::collections::HashMap;
use vrl::value::Kind;

use super::{
encoder::StackdriverLogsEncoder, request_builder::StackdriverLogsRequestBuilder,
service::StackdriverLogsServiceRequestBuilder, sink::StackdriverLogsSink,
};

#[derive(Debug, Snafu)]
enum HealthcheckError {
#[snafu(display("Resource not found"))]
NotFound,
}

/// Configuration for the `gcp_stackdriver_logs` sink.
#[configurable_component(sink(
"gcp_stackdriver_logs",
"Deliver logs to GCP's Cloud Operations suite."
))]
#[derive(Clone, Debug, Default)]
#[serde(deny_unknown_fields)]
pub(super) struct StackdriverConfig {
#[serde(skip, default = "default_endpoint")]
pub(super) endpoint: String,

#[serde(flatten)]
pub(super) log_name: StackdriverLogName,

/// The log ID to which to publish logs.
///
/// This is a name you create to identify this log stream.
pub(super) log_id: Template,

/// The monitored resource to associate the logs with.
pub(super) resource: StackdriverResource,

/// The field of the log event from which to take the outgoing log’s `severity` field.
///
/// The named field is removed from the log event if present, and must be either an integer
/// between 0 and 800 or a string containing one of the [severity level names][sev_names] (case
/// is ignored) or a common prefix such as `err`.
///
/// If no severity key is specified, the severity of outgoing records is set to 0 (`DEFAULT`).
///
/// See the [GCP Stackdriver Logging LogSeverity description][logsev_docs] for more details on
/// the value of the `severity` field.
///
/// [sev_names]: https://cloud.google.com/logging/docs/reference/v2/rest/v2/LogEntry#logseverity
/// [logsev_docs]: https://cloud.google.com/logging/docs/reference/v2/rest/v2/LogEntry#logseverity
#[configurable(metadata(docs::examples = "severity"))]
pub(super) severity_key: Option<ConfigValuePath>,

#[serde(flatten)]
pub(super) auth: GcpAuthConfig,

#[configurable(derived)]
#[serde(
default,
skip_serializing_if = "crate::serde::skip_serializing_if_default"
)]
pub(super) encoding: Transformer,

#[configurable(derived)]
#[serde(default)]
pub(super) batch: BatchConfig<RealtimeSizeBasedDefaultBatchSettings>,

#[configurable(derived)]
#[serde(default)]
pub(super) request: TowerRequestConfig,

#[configurable(derived)]
pub(super) tls: Option<TlsConfig>,

#[configurable(derived)]
#[serde(
default,
deserialize_with = "crate::serde::bool_or_struct",
skip_serializing_if = "crate::serde::skip_serializing_if_default"
)]
acknowledgements: AcknowledgementsConfig,
}

pub(super) fn default_endpoint() -> String {
"https://logging.googleapis.com/v2/entries:write".to_string()
}

// 10MB limit for entries.write: https://cloud.google.com/logging/quotas#api-limits
const MAX_BATCH_PAYLOAD_SIZE: usize = 10_000_000;

/// Logging locations.
#[configurable_component]
#[derive(Clone, Debug, Derivative)]
#[derivative(Default)]
pub(super) enum StackdriverLogName {
/// The billing account ID to which to publish logs.
///
/// Exactly one of `billing_account_id`, `folder_id`, `organization_id`, or `project_id` must be set.
#[serde(rename = "billing_account_id")]
#[configurable(metadata(docs::examples = "012345-6789AB-CDEF01"))]
BillingAccount(String),

/// The folder ID to which to publish logs.
///
/// See the [Google Cloud Platform folder documentation][folder_docs] for more details.
///
/// Exactly one of `billing_account_id`, `folder_id`, `organization_id`, or `project_id` must be set.
///
/// [folder_docs]: https://cloud.google.com/resource-manager/docs/creating-managing-folders
#[serde(rename = "folder_id")]
#[configurable(metadata(docs::examples = "My Folder"))]
Folder(String),

/// The organization ID to which to publish logs.
///
/// This would be the identifier assigned to your organization on Google Cloud Platform.
///
/// Exactly one of `billing_account_id`, `folder_id`, `organization_id`, or `project_id` must be set.
#[serde(rename = "organization_id")]
#[configurable(metadata(docs::examples = "622418129737"))]
Organization(String),

/// The project ID to which to publish logs.
///
/// See the [Google Cloud Platform project management documentation][project_docs] for more details.
///
/// Exactly one of `billing_account_id`, `folder_id`, `organization_id`, or `project_id` must be set.
///
/// [project_docs]: https://cloud.google.com/resource-manager/docs/creating-managing-projects
#[derivative(Default)]
#[serde(rename = "project_id")]
#[configurable(metadata(docs::examples = "vector-123456"))]
Project(String),
}

/// A monitored resource.
///
/// Monitored resources in GCP allow associating logs and metrics specifically with native resources
/// within Google Cloud Platform. This takes the form of a "type" field which identifies the
/// resource, and a set of type-specific labels to uniquely identify a resource of that type.
///
/// See [Monitored resource types][mon_docs] for more information.
///
/// [mon_docs]: https://cloud.google.com/monitoring/api/resources
// TODO: this type is specific to the stackdrivers log sink because it allows for template-able
// label values, but we should consider replacing `sinks::gcp::GcpTypedResource` with this so both
// the stackdriver metrics _and_ logs sink can have template-able label values, and less duplication
#[configurable_component]
#[derive(Clone, Debug, Default)]
pub(super) struct StackdriverResource {
/// The monitored resource type.
///
/// For example, the type of a Compute Engine VM instance is `gce_instance`.
/// See the [Google Cloud Platform monitored resource documentation][gcp_resources] for
/// more details.
///
/// [gcp_resources]: https://cloud.google.com/monitoring/api/resources
#[serde(rename = "type")]
pub(super) type_: String,

/// Type-specific labels.
#[serde(flatten)]
#[configurable(metadata(docs::additional_props_description = "A type-specific label."))]
#[configurable(metadata(docs::examples = "label_examples()"))]
pub(super) labels: HashMap<String, Template>,
}

fn label_examples() -> HashMap<String, String> {
let mut example = HashMap::new();
example.insert("instanceId".to_string(), "Twilight".to_string());
example.insert("zone".to_string(), "{{ zone }}".to_string());
example
}

impl_generate_config_from_default!(StackdriverConfig);

#[async_trait::async_trait]
#[typetag::serde(name = "gcp_stackdriver_logs")]
impl SinkConfig for StackdriverConfig {
async fn build(&self, cx: SinkContext) -> crate::Result<(VectorSink, Healthcheck)> {
let auth = self.auth.build(Scope::LoggingWrite).await?;

let request_builder = StackdriverLogsRequestBuilder {
encoder: StackdriverLogsEncoder::new(
self.encoding.clone(),
self.log_id.clone(),
self.log_name.clone(),
self.resource.clone(),
self.severity_key.clone(),
),
};

let batch_settings = self
.batch
.validate()?
.limit_max_bytes(MAX_BATCH_PAYLOAD_SIZE)?
.into_batcher_settings()?;

let request_limits = self.request.unwrap_with(
&TowerRequestConfig::default()
.rate_limit_duration_secs(1)
.rate_limit_num(1000),
);

let tls_settings = TlsSettings::from_options(&self.tls)?;
let client = HttpClient::new(tls_settings, cx.proxy())?;

let uri: Uri = self.endpoint.parse()?;

let stackdriver_logs_service_request_builder = StackdriverLogsServiceRequestBuilder {
uri: uri.clone(),
auth: auth.clone(),
};

let service = HttpService::new(client.clone(), stackdriver_logs_service_request_builder);

let service = ServiceBuilder::new()
.settings(request_limits, http_response_retry_logic())
.service(service);

let sink = StackdriverLogsSink::new(service, batch_settings, request_builder);

let healthcheck = healthcheck(client, auth.clone(), uri).boxed();

auth.spawn_regenerate_token();

Ok((VectorSink::from_event_streamsink(sink), healthcheck))
}

fn input(&self) -> Input {
let requirement =
schema::Requirement::empty().required_meaning("timestamp", Kind::timestamp());

Input::log().with_schema_requirement(requirement)
}

fn acknowledgements(&self) -> &AcknowledgementsConfig {
&self.acknowledgements
}
}

async fn healthcheck(client: HttpClient, auth: GcpAuthenticator, uri: Uri) -> crate::Result<()> {
let entries: Vec<BoxedRawValue> = Vec::new();
let events = serde_json::json!({ "entries": entries });

let body = crate::serde::json::to_bytes(&events).unwrap().freeze();

let mut request = Request::post(uri)
.header("Content-Type", "application/json")
.body(body)
.unwrap();

auth.apply(&mut request);

let request = request.map(Body::from);

let response = client.send(request).await?;

healthcheck_response(response, HealthcheckError::NotFound.into())
}
Loading