UNPKG

googleapis

Version:
802 lines 750 kB
/// <reference types="node" /> import { OAuth2Client, JWT, Compute, UserRefreshClient, BaseExternalAccountClient, GaxiosPromise, GoogleConfigurable, MethodOptions, StreamMethodOptions, GlobalOptions, GoogleAuth, BodyResponseCallback, APIRequestContext } from 'googleapis-common'; import { Readable } from 'stream'; export declare namespace logging_v2 { export interface Options extends GlobalOptions { version: 'v2'; } interface StandardParameters { /** * Auth client or API Key for the request */ auth?: string | OAuth2Client | JWT | Compute | UserRefreshClient | BaseExternalAccountClient | GoogleAuth; /** * V1 error format. */ '$.xgafv'?: string; /** * OAuth access token. */ access_token?: string; /** * Data format for response. */ alt?: string; /** * JSONP */ callback?: string; /** * Selector specifying which fields to include in a partial response. */ fields?: string; /** * API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token. */ key?: string; /** * OAuth 2.0 token for the current user. */ oauth_token?: string; /** * Returns response with indentations and line breaks. */ prettyPrint?: boolean; /** * Available to use for quota purposes for server-side applications. Can be any arbitrary string assigned to a user, but should not exceed 40 characters. */ quotaUser?: string; /** * Legacy upload protocol for media (e.g. "media", "multipart"). */ uploadType?: string; /** * Upload protocol for media (e.g. "raw", "multipart"). */ upload_protocol?: string; } /** * Cloud Logging API * * Writes log entries and manages your Cloud Logging configuration. The table entries below are presented in alphabetical order, not in order of common use. For explanations of the concepts found in the table entries, read the documentation at https://cloud.google.com/logging/docs. * * @example * ```js * const {google} = require('googleapis'); * const logging = google.logging('v2'); * ``` */ export class Logging { context: APIRequestContext; billingAccounts: Resource$Billingaccounts; entries: Resource$Entries; exclusions: Resource$Exclusions; folders: Resource$Folders; locations: Resource$Locations; logs: Resource$Logs; monitoredResourceDescriptors: Resource$Monitoredresourcedescriptors; organizations: Resource$Organizations; projects: Resource$Projects; sinks: Resource$Sinks; v2: Resource$V2; constructor(options: GlobalOptions, google?: GoogleConfigurable); } /** * Options that change functionality of a sink exporting data to BigQuery. */ export interface Schema$BigQueryOptions { /** * Optional. Whether to use BigQuery's partition tables (https://cloud.google.com/bigquery/docs/partitioned-tables). By default, Logging creates dated tables based on the log entries' timestamps, e.g. syslog_20170523. With partitioned tables the date suffix is no longer present and special query syntax (https://cloud.google.com/bigquery/docs/querying-partitioned-tables) has to be used instead. In both cases, tables are sharded based on UTC timezone. */ usePartitionedTables?: boolean | null; /** * Output only. True if new timestamp column based partitioning is in use, false if legacy ingestion-time partitioning is in use. All new sinks will have this field set true and will use timestamp column based partitioning. If use_partitioned_tables is false, this value has no meaning and will be false. Legacy sinks using partitioned tables will have this field set to false. */ usesTimestampColumnPartitioning?: boolean | null; } /** * BucketOptions describes the bucket boundaries used to create a histogram for the distribution. The buckets can be in a linear sequence, an exponential sequence, or each bucket can be specified explicitly. BucketOptions does not include the number of values in each bucket.A bucket has an inclusive lower bound and exclusive upper bound for the values that are counted for that bucket. The upper bound of a bucket must be strictly greater than the lower bound. The sequence of N buckets for a distribution consists of an underflow bucket (number 0), zero or more finite buckets (number 1 through N - 2) and an overflow bucket (number N - 1). The buckets are contiguous: the lower bound of bucket i (i \> 0) is the same as the upper bound of bucket i - 1. The buckets span the whole range of finite values: lower bound of the underflow bucket is -infinity and the upper bound of the overflow bucket is +infinity. The finite buckets are so-called because both bounds are finite. */ export interface Schema$BucketOptions { /** * The explicit buckets. */ explicitBuckets?: Schema$Explicit; /** * The exponential buckets. */ exponentialBuckets?: Schema$Exponential; /** * The linear bucket. */ linearBuckets?: Schema$Linear; } /** * Describes the customer-managed encryption key (CMEK) settings associated with a project, folder, organization, billing account, or flexible resource.Note: CMEK for the Logs Router can currently only be configured for GCP organizations. Once configured, it applies to all projects and folders in the GCP organization.See Enabling CMEK for Logs Router (https://cloud.google.com/logging/docs/routing/managed-encryption) for more information. */ export interface Schema$CmekSettings { /** * The resource name for the configured Cloud KMS key.KMS key name format: "projects/PROJECT_ID/locations/LOCATION/keyRings/KEYRING/cryptoKeys/KEY"For example: "projects/my-project-id/locations/my-region/keyRings/key-ring-name/cryptoKeys/key-name"To enable CMEK for the Logs Router, set this field to a valid kms_key_name for which the associated service account has the required roles/cloudkms.cryptoKeyEncrypterDecrypter role assigned for the key.The Cloud KMS key used by the Log Router can be updated by changing the kms_key_name to a new valid key name. Encryption operations that are in progress will be completed with the key that was in use when they started. Decryption operations will be completed using the key that was used at the time of encryption unless access to that key has been revoked.To disable CMEK for the Logs Router, set this field to an empty string.See Enabling CMEK for Logs Router (https://cloud.google.com/logging/docs/routing/managed-encryption) for more information. */ kmsKeyName?: string | null; /** * Output only. The resource name of the CMEK settings. */ name?: string | null; /** * Output only. The service account that will be used by the Logs Router to access your Cloud KMS key.Before enabling CMEK for Logs Router, you must first assign the role roles/cloudkms.cryptoKeyEncrypterDecrypter to the service account that the Logs Router will use to access your Cloud KMS key. Use GetCmekSettings to obtain the service account ID.See Enabling CMEK for Logs Router (https://cloud.google.com/logging/docs/routing/managed-encryption) for more information. */ serviceAccountId?: string | null; } /** * A generic empty message that you can re-use to avoid defining duplicated empty messages in your APIs. A typical example is to use it as the request or the response type of an API method. For instance: service Foo { rpc Bar(google.protobuf.Empty) returns (google.protobuf.Empty); \} The JSON representation for Empty is empty JSON object {\}. */ export interface Schema$Empty { } /** * Specifies a set of buckets with arbitrary widths.There are size(bounds) + 1 (= N) buckets. Bucket i has the following boundaries:Upper bound (0 <= i < N-1): boundsi Lower bound (1 <= i < N); boundsi - 1The bounds field must contain at least one element. If bounds has only one element, then there are no finite buckets, and that single element is the common boundary of the overflow and underflow buckets. */ export interface Schema$Explicit { /** * The values must be monotonically increasing. */ bounds?: number[] | null; } /** * Specifies an exponential sequence of buckets that have a width that is proportional to the value of the lower bound. Each bucket represents a constant relative uncertainty on a specific value in the bucket.There are num_finite_buckets + 2 (= N) buckets. Bucket i has the following boundaries:Upper bound (0 <= i < N-1): scale * (growth_factor ^ i). Lower bound (1 <= i < N): scale * (growth_factor ^ (i - 1)). */ export interface Schema$Exponential { /** * Must be greater than 1. */ growthFactor?: number | null; /** * Must be greater than 0. */ numFiniteBuckets?: number | null; /** * Must be greater than 0. */ scale?: number | null; } /** * A common proto for logging HTTP requests. Only contains semantics defined by the HTTP specification. Product-specific logging information MUST be defined in a separate message. */ export interface Schema$HttpRequest { /** * The number of HTTP response bytes inserted into cache. Set only when a cache fill was attempted. */ cacheFillBytes?: string | null; /** * Whether or not an entity was served from cache (with or without validation). */ cacheHit?: boolean | null; /** * Whether or not a cache lookup was attempted. */ cacheLookup?: boolean | null; /** * Whether or not the response was validated with the origin server before being served from cache. This field is only meaningful if cache_hit is True. */ cacheValidatedWithOriginServer?: boolean | null; /** * The request processing latency on the server, from the time the request was received until the response was sent. */ latency?: string | null; /** * Protocol used for the request. Examples: "HTTP/1.1", "HTTP/2", "websocket" */ protocol?: string | null; /** * The referer URL of the request, as defined in HTTP/1.1 Header Field Definitions (http://www.w3.org/Protocols/rfc2616/rfc2616-sec14.html). */ referer?: string | null; /** * The IP address (IPv4 or IPv6) of the client that issued the HTTP request. This field can include port information. Examples: "192.168.1.1", "10.0.0.1:80", "FE80::0202:B3FF:FE1E:8329". */ remoteIp?: string | null; /** * The request method. Examples: "GET", "HEAD", "PUT", "POST". */ requestMethod?: string | null; /** * The size of the HTTP request message in bytes, including the request headers and the request body. */ requestSize?: string | null; /** * The scheme (http, https), the host name, the path and the query portion of the URL that was requested. Example: "http://example.com/some/info?color=red". */ requestUrl?: string | null; /** * The size of the HTTP response message sent back to the client, in bytes, including the response headers and the response body. */ responseSize?: string | null; /** * The IP address (IPv4 or IPv6) of the origin server that the request was sent to. This field can include port information. Examples: "192.168.1.1", "10.0.0.1:80", "FE80::0202:B3FF:FE1E:8329". */ serverIp?: string | null; /** * The response code indicating the status of response. Examples: 200, 404. */ status?: number | null; /** * The user agent sent by the client. Example: "Mozilla/4.0 (compatible; MSIE 6.0; Windows 98; Q312461; .NET CLR 1.0.3705)". */ userAgent?: string | null; } /** * A description of a label. */ export interface Schema$LabelDescriptor { /** * A human-readable description for the label. */ description?: string | null; /** * The label key. */ key?: string | null; /** * The type of data that can be assigned to the label. */ valueType?: string | null; } /** * Specifies a linear sequence of buckets that all have the same width (except overflow and underflow). Each bucket represents a constant absolute uncertainty on the specific value in the bucket.There are num_finite_buckets + 2 (= N) buckets. Bucket i has the following boundaries:Upper bound (0 <= i < N-1): offset + (width * i). Lower bound (1 <= i < N): offset + (width * (i - 1)). */ export interface Schema$Linear { /** * Must be greater than 0. */ numFiniteBuckets?: number | null; /** * Lower bound of the first bucket. */ offset?: number | null; /** * Must be greater than 0. */ width?: number | null; } /** * The response from ListBuckets. */ export interface Schema$ListBucketsResponse { /** * A list of buckets. */ buckets?: Schema$LogBucket[]; /** * If there might be more results than appear in this response, then nextPageToken is included. To get the next set of results, call the same method again using the value of nextPageToken as pageToken. */ nextPageToken?: string | null; } /** * Result returned from ListExclusions. */ export interface Schema$ListExclusionsResponse { /** * A list of exclusions. */ exclusions?: Schema$LogExclusion[]; /** * If there might be more results than appear in this response, then nextPageToken is included. To get the next set of results, call the same method again using the value of nextPageToken as pageToken. */ nextPageToken?: string | null; } /** * The response message for Locations.ListLocations. */ export interface Schema$ListLocationsResponse { /** * A list of locations that matches the specified filter in the request. */ locations?: Schema$Location[]; /** * The standard List next-page token. */ nextPageToken?: string | null; } /** * The parameters to ListLogEntries. */ export interface Schema$ListLogEntriesRequest { /** * Optional. A filter that chooses which log entries to return. See Advanced Logs Queries (https://cloud.google.com/logging/docs/view/advanced-queries). Only log entries that match the filter are returned. An empty filter matches all log entries in the resources listed in resource_names. Referencing a parent resource that is not listed in resource_names will cause the filter to return no results. The maximum length of the filter is 20000 characters. */ filter?: string | null; /** * Optional. How the results should be sorted. Presently, the only permitted values are "timestamp asc" (default) and "timestamp desc". The first option returns entries in order of increasing values of LogEntry.timestamp (oldest first), and the second option returns entries in order of decreasing timestamps (newest first). Entries with equal timestamps are returned in order of their insert_id values. */ orderBy?: string | null; /** * Optional. The maximum number of results to return from this request. Default is 50. If the value is negative or exceeds 1000, the request is rejected. The presence of next_page_token in the response indicates that more results might be available. */ pageSize?: number | null; /** * Optional. If present, then retrieve the next batch of results from the preceding call to this method. page_token must be the value of next_page_token from the previous response. The values of other method parameters should be identical to those in the previous call. */ pageToken?: string | null; /** * Optional. Deprecated. Use resource_names instead. One or more project identifiers or project numbers from which to retrieve log entries. Example: "my-project-1A". */ projectIds?: string[] | null; /** * Required. Names of one or more parent resources from which to retrieve log entries: "projects/[PROJECT_ID]" "organizations/[ORGANIZATION_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]" "folders/[FOLDER_ID]" May alternatively be one or more views projects/PROJECT_ID/locations/LOCATION_ID/buckets/BUCKET_ID/views/VIEW_ID organization/ORGANIZATION_ID/locations/LOCATION_ID/buckets/BUCKET_ID/views/VIEW_ID billingAccounts/BILLING_ACCOUNT_ID/locations/LOCATION_ID/buckets/BUCKET_ID/views/VIEW_ID folders/FOLDER_ID/locations/LOCATION_ID/buckets/BUCKET_ID/views/VIEW_IDProjects listed in the project_ids field are added to this list. */ resourceNames?: string[] | null; } /** * Result returned from ListLogEntries. */ export interface Schema$ListLogEntriesResponse { /** * A list of log entries. If entries is empty, nextPageToken may still be returned, indicating that more entries may exist. See nextPageToken for more information. */ entries?: Schema$LogEntry[]; /** * If there might be more results than those appearing in this response, then nextPageToken is included. To get the next set of results, call this method again using the value of nextPageToken as pageToken.If a value for next_page_token appears and the entries field is empty, it means that the search found no log entries so far but it did not have time to search all the possible log entries. Retry the method with this value for page_token to continue the search. Alternatively, consider speeding up the search by changing your filter to specify a single log name or resource type, or to narrow the time range of the search. */ nextPageToken?: string | null; } /** * Result returned from ListLogMetrics. */ export interface Schema$ListLogMetricsResponse { /** * A list of logs-based metrics. */ metrics?: Schema$LogMetric[]; /** * If there might be more results than appear in this response, then nextPageToken is included. To get the next set of results, call this method again using the value of nextPageToken as pageToken. */ nextPageToken?: string | null; } /** * Result returned from ListLogs. */ export interface Schema$ListLogsResponse { /** * A list of log names. For example, "projects/my-project/logs/syslog" or "organizations/123/logs/cloudresourcemanager.googleapis.com%2Factivity". */ logNames?: string[] | null; /** * If there might be more results than those appearing in this response, then nextPageToken is included. To get the next set of results, call this method again using the value of nextPageToken as pageToken. */ nextPageToken?: string | null; } /** * Result returned from ListMonitoredResourceDescriptors. */ export interface Schema$ListMonitoredResourceDescriptorsResponse { /** * If there might be more results than those appearing in this response, then nextPageToken is included. To get the next set of results, call this method again using the value of nextPageToken as pageToken. */ nextPageToken?: string | null; /** * A list of resource descriptors. */ resourceDescriptors?: Schema$MonitoredResourceDescriptor[]; } /** * Result returned from ListSinks. */ export interface Schema$ListSinksResponse { /** * If there might be more results than appear in this response, then nextPageToken is included. To get the next set of results, call the same method again using the value of nextPageToken as pageToken. */ nextPageToken?: string | null; /** * A list of sinks. */ sinks?: Schema$LogSink[]; } /** * The response from ListViews. */ export interface Schema$ListViewsResponse { /** * If there might be more results than appear in this response, then nextPageToken is included. To get the next set of results, call the same method again using the value of nextPageToken as pageToken. */ nextPageToken?: string | null; /** * A list of views. */ views?: Schema$LogView[]; } /** * A resource that represents Google Cloud Platform location. */ export interface Schema$Location { /** * The friendly name for this location, typically a nearby city name. For example, "Tokyo". */ displayName?: string | null; /** * Cross-service attributes for the location. For example {"cloud.googleapis.com/region": "us-east1"\} */ labels?: { [key: string]: string; } | null; /** * The canonical id for this location. For example: "us-east1". */ locationId?: string | null; /** * Service-specific metadata. For example the available capacity at the given location. */ metadata?: { [key: string]: any; } | null; /** * Resource name for the location, which may vary between implementations. For example: "projects/example-project/locations/us-east1" */ name?: string | null; } /** * Describes a repository of logs. */ export interface Schema$LogBucket { /** * Output only. The creation timestamp of the bucket. This is not set for any of the default buckets. */ createTime?: string | null; /** * Describes this bucket. */ description?: string | null; /** * Output only. The bucket lifecycle state. */ lifecycleState?: string | null; /** * Whether the bucket has been locked. The retention period on a locked bucket may not be changed. Locked buckets may only be deleted if they are empty. */ locked?: boolean | null; /** * Output only. The resource name of the bucket. For example: "projects/my-project-id/locations/my-location/buckets/my-bucket-id" The supported locations are: global, us-central1, us-east1, us-west1, asia-east1, europe-west1.For the location of global it is unspecified where logs are actually stored. Once a bucket has been created, the location can not be changed. */ name?: string | null; /** * Logs will be retained by default for this amount of time, after which they will automatically be deleted. The minimum retention period is 1 day. If this value is set to zero at bucket creation time, the default time of 30 days will be used. */ retentionDays?: number | null; /** * Output only. The last update timestamp of the bucket. */ updateTime?: string | null; } /** * An individual entry in a log. */ export interface Schema$LogEntry { /** * Optional. Information about the HTTP request associated with this log entry, if applicable. */ httpRequest?: Schema$HttpRequest; /** * Optional. A unique identifier for the log entry. If you provide a value, then Logging considers other log entries in the same project, with the same timestamp, and with the same insert_id to be duplicates which are removed in a single query result. However, there are no guarantees of de-duplication in the export of logs.If the insert_id is omitted when writing a log entry, the Logging API assigns its own unique identifier in this field.In queries, the insert_id is also used to order log entries that have the same log_name and timestamp values. */ insertId?: string | null; /** * The log entry payload, represented as a structure that is expressed as a JSON object. */ jsonPayload?: { [key: string]: any; } | null; /** * Optional. A set of user-defined (key, value) data that provides additional information about the log entry.Cloud Logging truncates label keys that exceed 512 B and label values that exceed 64 KB upon their associated log entry being written. The truncation is indicated by an ellipsis at the end of the character string. */ labels?: { [key: string]: string; } | null; /** * Required. The resource name of the log to which this log entry belongs: "projects/[PROJECT_ID]/logs/[LOG_ID]" "organizations/[ORGANIZATION_ID]/logs/[LOG_ID]" "billingAccounts/[BILLING_ACCOUNT_ID]/logs/[LOG_ID]" "folders/[FOLDER_ID]/logs/[LOG_ID]" A project number may be used in place of PROJECT_ID. The project number is translated to its corresponding PROJECT_ID internally and the log_name field will contain PROJECT_ID in queries and exports.[LOG_ID] must be URL-encoded within log_name. Example: "organizations/1234567890/logs/cloudresourcemanager.googleapis.com%2Factivity". [LOG_ID] must be less than 512 characters long and can only include the following characters: upper and lower case alphanumeric characters, forward-slash, underscore, hyphen, and period.For backward compatibility, if log_name begins with a forward-slash, such as /projects/..., then the log entry is ingested as usual but the forward-slash is removed. Listing the log entry will not show the leading slash and filtering for a log name with a leading slash will never return any results. */ logName?: string | null; /** * Output only. Deprecated. Additional metadata about the monitored resource.Only k8s_container, k8s_pod, and k8s_node MonitoredResources have this field populated for GKE versions older than 1.12.6. For GKE versions 1.12.6 and above, the metadata field has been deprecated. The Kubernetes pod labels that used to be in metadata.userLabels will now be present in the labels field with a key prefix of k8s-pod/. The system labels that were present in the metadata.systemLabels field will no longer be available in the log entry. */ metadata?: Schema$MonitoredResourceMetadata; /** * Optional. Information about an operation associated with the log entry, if applicable. */ operation?: Schema$LogEntryOperation; /** * The log entry payload, represented as a protocol buffer. Some Google Cloud Platform services use this field for their log entry payloads.The following protocol buffer types are supported; user-defined types are not supported:"type.googleapis.com/google.cloud.audit.AuditLog" "type.googleapis.com/google.appengine.logging.v1.RequestLog" */ protoPayload?: { [key: string]: any; } | null; /** * Output only. The time the log entry was received by Logging. */ receiveTimestamp?: string | null; /** * Required. The monitored resource that produced this log entry.Example: a log entry that reports a database error would be associated with the monitored resource designating the particular database that reported the error. */ resource?: Schema$MonitoredResource; /** * Optional. The severity of the log entry. The default value is LogSeverity.DEFAULT. */ severity?: string | null; /** * Optional. Source code location information associated with the log entry, if any. */ sourceLocation?: Schema$LogEntrySourceLocation; /** * Optional. The span ID within the trace associated with the log entry.For Trace spans, this is the same format that the Trace API v2 uses: a 16-character hexadecimal encoding of an 8-byte array, such as 000000000000004a. */ spanId?: string | null; /** * The log entry payload, represented as a Unicode string (UTF-8). */ textPayload?: string | null; /** * Optional. The time the event described by the log entry occurred. This time is used to compute the log entry's age and to enforce the logs retention period. If this field is omitted in a new log entry, then Logging assigns it the current time. Timestamps have nanosecond accuracy, but trailing zeros in the fractional seconds might be omitted when the timestamp is displayed.Incoming log entries must have timestamps that don't exceed the logs retention period (https://cloud.google.com/logging/quotas#logs_retention_periods) in the past, and that don't exceed 24 hours in the future. Log entries outside those time boundaries aren't ingested by Logging. */ timestamp?: string | null; /** * Optional. Resource name of the trace associated with the log entry, if any. If it contains a relative resource name, the name is assumed to be relative to //tracing.googleapis.com. Example: projects/my-projectid/traces/06796866738c859f2f19b7cfb3214824 */ trace?: string | null; /** * Optional. The sampling decision of the trace associated with the log entry.True means that the trace resource name in the trace field was sampled for storage in a trace backend. False means that the trace was not sampled for storage when this log entry was written, or the sampling decision was unknown at the time. A non-sampled trace value is still useful as a request correlation identifier. The default is False. */ traceSampled?: boolean | null; } /** * Additional information about a potentially long-running operation with which a log entry is associated. */ export interface Schema$LogEntryOperation { /** * Optional. Set this to True if this is the first log entry in the operation. */ first?: boolean | null; /** * Optional. An arbitrary operation identifier. Log entries with the same identifier are assumed to be part of the same operation. */ id?: string | null; /** * Optional. Set this to True if this is the last log entry in the operation. */ last?: boolean | null; /** * Optional. An arbitrary producer identifier. The combination of id and producer must be globally unique. Examples for producer: "MyDivision.MyBigCompany.com", "github.com/MyProject/MyApplication". */ producer?: string | null; } /** * Additional information about the source code location that produced the log entry. */ export interface Schema$LogEntrySourceLocation { /** * Optional. Source file name. Depending on the runtime environment, this might be a simple name or a fully-qualified name. */ file?: string | null; /** * Optional. Human-readable name of the function or method being invoked, with optional context such as the class or package name. This information may be used in contexts such as the logs viewer, where a file and line number are less meaningful. The format can vary by language. For example: qual.if.ied.Class.method (Java), dir/package.func (Go), function (Python). */ function?: string | null; /** * Optional. Line within the source file. 1-based; 0 indicates no line number available. */ line?: string | null; } /** * Specifies a set of log entries that are not to be stored in Logging. If your GCP resource receives a large volume of logs, you can use exclusions to reduce your chargeable logs. Exclusions are processed after log sinks, so you can export log entries before they are excluded. Note that organization-level and folder-level exclusions don't apply to child resources, and that you can't exclude audit log entries. */ export interface Schema$LogExclusion { /** * Output only. The creation timestamp of the exclusion.This field may not be present for older exclusions. */ createTime?: string | null; /** * Optional. A description of this exclusion. */ description?: string | null; /** * Optional. If set to True, then this exclusion is disabled and it does not exclude any log entries. You can update an exclusion to change the value of this field. */ disabled?: boolean | null; /** * Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries) that matches the log entries to be excluded. By using the sample function (https://cloud.google.com/logging/docs/view/advanced-queries#sample), you can exclude less than 100% of the matching log entries. For example, the following query matches 99% of low-severity log entries from Google Cloud Storage buckets:"resource.type=gcs_bucket severity<ERROR sample(insertId, 0.99)" */ filter?: string | null; /** * Required. A client-assigned identifier, such as "load-balancer-exclusion". Identifiers are limited to 100 characters and can include only letters, digits, underscores, hyphens, and periods. First character has to be alphanumeric. */ name?: string | null; /** * Output only. The last update timestamp of the exclusion.This field may not be present for older exclusions. */ updateTime?: string | null; } /** * Application log line emitted while processing a request. */ export interface Schema$LogLine { /** * App-provided log message. */ logMessage?: string | null; /** * Severity of this log entry. */ severity?: string | null; /** * Where in the source code this log message was written. */ sourceLocation?: Schema$SourceLocation; /** * Approximate time when this log entry was made. */ time?: string | null; } /** * Describes a logs-based metric. The value of the metric is the number of log entries that match a logs filter in a given time interval.Logs-based metrics can also be used to extract values from logs and create a distribution of the values. The distribution records the statistics of the extracted values along with an optional histogram of the values as specified by the bucket options. */ export interface Schema$LogMetric { /** * Optional. The bucket_options are required when the logs-based metric is using a DISTRIBUTION value type and it describes the bucket boundaries used to create a histogram of the extracted values. */ bucketOptions?: Schema$BucketOptions; /** * Output only. The creation timestamp of the metric.This field may not be present for older metrics. */ createTime?: string | null; /** * Optional. A description of this metric, which is used in documentation. The maximum length of the description is 8000 characters. */ description?: string | null; /** * Required. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced_filters) which is used to match log entries. Example: "resource.type=gae_app AND severity\>=ERROR" The maximum length of the filter is 20000 characters. */ filter?: string | null; /** * Optional. A map from a label key string to an extractor expression which is used to extract data from a log entry field and assign as the label value. Each label key specified in the LabelDescriptor must have an associated extractor expression in this map. The syntax of the extractor expression is the same as for the value_extractor field.The extracted value is converted to the type defined in the label descriptor. If the either the extraction or the type conversion fails, the label will have a default value. The default value for a string label is an empty string, for an integer label its 0, and for a boolean label its false.Note that there are upper bounds on the maximum number of labels and the number of active time series that are allowed in a project. */ labelExtractors?: { [key: string]: string; } | null; /** * Optional. The metric descriptor associated with the logs-based metric. If unspecified, it uses a default metric descriptor with a DELTA metric kind, INT64 value type, with no labels and a unit of "1". Such a metric counts the number of log entries matching the filter expression.The name, type, and description fields in the metric_descriptor are output only, and is constructed using the name and description field in the LogMetric.To create a logs-based metric that records a distribution of log values, a DELTA metric kind with a DISTRIBUTION value type must be used along with a value_extractor expression in the LogMetric.Each label in the metric descriptor must have a matching label name as the key and an extractor expression as the value in the label_extractors map.The metric_kind and value_type fields in the metric_descriptor cannot be updated once initially configured. New labels can be added in the metric_descriptor, but existing labels cannot be modified except for their description. */ metricDescriptor?: Schema$MetricDescriptor; /** * Required. The client-assigned metric identifier. Examples: "error_count", "nginx/requests".Metric identifiers are limited to 100 characters and can include only the following characters: A-Z, a-z, 0-9, and the special characters _-.,+!*',()%/. The forward-slash character (/) denotes a hierarchy of name pieces, and it cannot be the first character of the name.The metric identifier in this field must not be URL-encoded (https://en.wikipedia.org/wiki/Percent-encoding). However, when the metric identifier appears as the [METRIC_ID] part of a metric_name API parameter, then the metric identifier must be URL-encoded. Example: "projects/my-project/metrics/nginx%2Frequests". */ name?: string | null; /** * Output only. The last update timestamp of the metric.This field may not be present for older metrics. */ updateTime?: string | null; /** * Optional. A value_extractor is required when using a distribution logs-based metric to extract the values to record from a log entry. Two functions are supported for value extraction: EXTRACT(field) or REGEXP_EXTRACT(field, regex). The argument are: 1. field: The name of the log entry field from which the value is to be extracted. 2. regex: A regular expression using the Google RE2 syntax (https://github.com/google/re2/wiki/Syntax) with a single capture group to extract data from the specified log entry field. The value of the field is converted to a string before applying the regex. It is an error to specify a regex that does not include exactly one capture group.The result of the extraction must be convertible to a double type, as the distribution always records double values. If either the extraction or the conversion to double fails, then those values are not recorded in the distribution.Example: REGEXP_EXTRACT(jsonPayload.request, ".*quantity=(\d+).*") */ valueExtractor?: string | null; /** * Deprecated. The API version that created or updated this metric. The v2 format is used by default and cannot be changed. */ version?: string | null; } /** * Describes a sink used to export log entries to one of the following destinations in any project: a Cloud Storage bucket, a BigQuery dataset, a Cloud Pub/Sub topic or a Cloud Logging Bucket. A logs filter controls which log entries are exported. The sink must be created within a project, organization, billing account, or folder. */ export interface Schema$LogSink { /** * Optional. Options that affect sinks exporting data to BigQuery. */ bigqueryOptions?: Schema$BigQueryOptions; /** * Output only. The creation timestamp of the sink.This field may not be present for older sinks. */ createTime?: string | null; /** * Optional. A description of this sink. The maximum length of the description is 8000 characters. */ description?: string | null; /** * Required. The export destination: "storage.googleapis.com/[GCS_BUCKET]" "bigquery.googleapis.com/projects/[PROJECT_ID]/datasets/[DATASET]" "pubsub.googleapis.com/projects/[PROJECT_ID]/topics/[TOPIC_ID]" The sink's writer_identity, set when the sink is created, must have permission to write to the destination or else the log entries are not exported. For more information, see Exporting Logs with Sinks (https://cloud.google.com/logging/docs/api/tasks/exporting-logs). */ destination?: string | null; /** * Optional. If set to True, then this sink is disabled and it does not export any log entries. */ disabled?: boolean | null; /** * Optional. Log entries that match any of the exclusion filters will not be exported. If a log entry is matched by both filter and one of exclusion_filters it will not be exported. */ exclusions?: Schema$LogExclusion[]; /** * Optional. An advanced logs filter (https://cloud.google.com/logging/docs/view/advanced-queries). The only exported log entries are those that are in the resource owning the sink and that match the filter. For example: logName="projects/[PROJECT_ID]/logs/[LOG_ID]" AND severity\>=ERROR */ filter?: string | null; /** * Optional. This field applies only to sinks owned by organizations and folders. If the field is false, the default, only the logs owned by the sink's parent resource are available for export. If the field is true, then logs from all the projects, folders, and billing accounts contained in the sink's parent resource are also available for export. Whether a particular log entry from the children is exported depends on the sink's filter expression. For example, if this field is true, then the filter resource.type=gce_instance would export all Compute Engine VM instance log entries from all projects in the sink's parent. To only export entries from certain child projects, filter on the project part of the log name: logName:("projects/test-project1/" OR "projects/test-project2/") AND resource.type=gce_instance */ includeChildren?: boolean | null; /** * Required. The client-assigned sink identifier, unique within the project. Example: "my-syslog-errors-to-pubsub". Sink identifiers are limited to 100 characters and can include only the following characters: upper and lower-case alphanumeric characters, underscores, hyphens, and periods. First character has to be alphanumeric. */ name?: string | null; /** * Deprecated. This field is unused. */ outputVersionFormat?: string | null; /** * Output only. The last update timestamp of the sink.This field may not be present for older sinks. */ updateTime?: string | null; /** * Output only. An IAM identity—a service account or group—under which Logging writes the exported log entries to the sink's destination. This field is set by sinks.create and sinks.update based on the value of unique_writer_identity in those methods.Until you grant this identity write-access to the destination, log entry exports from this sink will fail. For more information, see Granting Access for a Resource (https://cloud.google.com/iam/docs/granting-roles-to-service-accounts#granting_access_to_a_service_account_for_a_resource). Consult the destination service's documentation to determine the appropriate IAM roles to assign to the identity. */ writerIdentity?: string | null; } /** * Describes a view over logs in a bucket. */ export interface Schema$LogView { /** * Output only. The creation timestamp of the view. */ createTime?: string | null; /** * Describes this view. */ description?: string | null; /** * Filter that restricts which log entries in a bucket are visible in this view. Filters are restricted to be a logical AND of ==/!= of any of the following: originating project/folder/organization/billing account. resource type log id Example: SOURCE("projects/myproject") AND resource.type = "gce_instance" AND LOG_ID("stdout") */ filter?: string | null; /** * The resource name of the view. For example "projects/my-project-id/locations/my-location/buckets/my-bucket-id/views/my-view */ name?: string | null; /** * Output only. The last update timestamp of the view. */ updateTime?: string | null; } /** * Defines a metric type and its schema. Once a metric descriptor is created, deleting or altering it stops data collection and makes the metric type's existing data unusable. */ export interface Schema$MetricDescriptor { /** * A detailed description of the metric, which can be used in documentation. */ description?: string | null; /** * A concise name for the metric, which can be displayed in user interfaces. Use sentence case without an ending period, for example "Request count". This field is optional but it is recommended to be set for any metrics associated with user-visible concepts, such as Quota. */ displayName?: string | null; /** * The set of labels that can be used to describe a specific instance of this metric type. For example, the appengine.googleapis.com/http/server/response_latencies metric type has a label for the HTTP response code, response_code, so you can look at latencies for successful responses or just for responses that failed. */ labels?: Schema$LabelDescriptor[]; /** * Optional. The launch stage of the metric definition. */ launchStage?: string | null; /** * Optional. Metadata which can be used to guide usage of the metric. */ metadata?: Schema$MetricDescriptorMetadata; /** * Whether the metric records instantaneous values, changes to a value, etc. Some combinations of metric_kind and value_type might not be supported. */ metricKind?: string | null; /** * Read-only. If present, then a time series, which is identified partially by a metric type and a MonitoredResourceDescriptor, that is associated with this metric type can only be associated with one of the monitored resource types listed here. */ monitoredResourceTypes?: string[] | null; /** * The resource name of the metric descriptor. */ name?: string | null; /** * The metric type, including its DNS name prefix. The type is not URL-encoded. All user-defined metric types have the DNS name custom.googleapis.com or external.googleapis.com. Metric types should use a natural hierarchical grouping. For example: "custom.googleapis.com/invoice/paid/amount" "external.googleapis.com/prometheus/up" "appengine.googleapis.com/http/server/response_latencies" */ type?: string | null; /** * The units in which the metric value is reported. It is only applicable if the value_type is INT64, DOUBLE, or DISTRIBUTION. The unit defines the representation of the stored metric values.Different systems might scale the values to be more easily displayed (so a value of 0.02kBy might be displayed as 20By, and a value of 3523kBy might be displayed as 3.5MBy). However, if the unit is kBy, then the value of the metric is always in thousands of bytes, no matter how it might be displayed.If you want a custom metric to record the exact number of CPU-seconds used by a job, you can create an INT64 CUMULATIVE metric whose unit is s{CPU\} (or equivalently 1s{CPU\} or just s). If the job uses 12,005 CPU-seconds, then the value is written as 12005.Alternatively, if you want a custom metric to record data in a more granular way, you can create a DOUBLE CUMULATIVE metric whose unit is ks{CPU\}, and then write the value 12.005 (which is 12005/1000), or use Kis{CPU\} and write 11.723 (which is 12005/1024).The supported units are a subset of The Unified Code for Units of Measure (https://unitsofmeasure.org/ucum.html) standard:Basic units (UNIT) bit bit By byte s second min minute h hour d day 1 dimensionlessPrefixes (PREFIX) k kilo (10^3) M mega (10^6) G giga (10^9) T tera (10^12) P peta (10^15) E exa (10^18) Z zetta (10^21) Y yotta (10^24) m mill