googleapis
Version:
Google APIs Client Library for Node.js
1,017 lines (1,016 loc) • 279 kB
TypeScript
/// <reference types="node" />
import { OAuth2Client, JWT, Compute, UserRefreshClient, BaseExternalAccountClient, GaxiosPromise, GoogleConfigurable, MethodOptions, StreamMethodOptions, GlobalOptions, GoogleAuth, BodyResponseCallback, APIRequestContext } from 'googleapis-common';
import { Readable } from 'stream';
export declare namespace bigquery_v2 {
export interface Options extends GlobalOptions {
version: 'v2';
}
interface StandardParameters {
/**
* Auth client or API Key for the request
*/
auth?: string | OAuth2Client | JWT | Compute | UserRefreshClient | BaseExternalAccountClient | GoogleAuth;
/**
* Data format for the response.
*/
alt?: string;
/**
* Selector specifying which fields to include in a partial response.
*/
fields?: string;
/**
* API key. Your API key identifies your project and provides you with API access, quota, and reports. Required unless you provide an OAuth 2.0 token.
*/
key?: string;
/**
* OAuth 2.0 token for the current user.
*/
oauth_token?: string;
/**
* Returns response with indentations and line breaks.
*/
prettyPrint?: boolean;
/**
* An opaque string that represents a user for quota purposes. Must not exceed 40 characters.
*/
quotaUser?: string;
/**
* Deprecated. Please use quotaUser instead.
*/
userIp?: string;
}
/**
* BigQuery API
*
* A data platform for customers to create, manage, share and query data.
*
* @example
* ```js
* const {google} = require('googleapis');
* const bigquery = google.bigquery('v2');
* ```
*/
export class Bigquery {
context: APIRequestContext;
datasets: Resource$Datasets;
jobs: Resource$Jobs;
models: Resource$Models;
projects: Resource$Projects;
routines: Resource$Routines;
rowAccessPolicies: Resource$Rowaccesspolicies;
tabledata: Resource$Tabledata;
tables: Resource$Tables;
constructor(options: GlobalOptions, google?: GoogleConfigurable);
}
/**
* Aggregate metrics for classification/classifier models. For multi-class models, the metrics are either macro-averaged or micro-averaged. When macro-averaged, the metrics are calculated for each label and then an unweighted average is taken of those values. When micro-averaged, the metric is calculated globally by counting the total number of correctly predicted rows.
*/
export interface Schema$AggregateClassificationMetrics {
/**
* Accuracy is the fraction of predictions given the correct label. For multiclass this is a micro-averaged metric.
*/
accuracy?: number | null;
/**
* The F1 score is an average of recall and precision. For multiclass this is a macro-averaged metric.
*/
f1Score?: number | null;
/**
* Logarithmic Loss. For multiclass this is a macro-averaged metric.
*/
logLoss?: number | null;
/**
* Precision is the fraction of actual positive predictions that had positive actual labels. For multiclass this is a macro-averaged metric treating each class as a binary classifier.
*/
precision?: number | null;
/**
* Recall is the fraction of actual positive labels that were given a positive prediction. For multiclass this is a macro-averaged metric.
*/
recall?: number | null;
/**
* Area Under a ROC Curve. For multiclass this is a macro-averaged metric.
*/
rocAuc?: number | null;
/**
* Threshold at which the metrics are computed. For binary classification models this is the positive class threshold. For multi-class classfication models this is the confidence threshold.
*/
threshold?: number | null;
}
/**
* Input/output argument of a function or a stored procedure.
*/
export interface Schema$Argument {
/**
* Optional. Defaults to FIXED_TYPE.
*/
argumentKind?: string | null;
/**
* Required unless argument_kind = ANY_TYPE.
*/
dataType?: Schema$StandardSqlDataType;
/**
* Optional. Specifies whether the argument is input or output. Can be set for procedures only.
*/
mode?: string | null;
/**
* Optional. The name of this argument. Can be absent for function return argument.
*/
name?: string | null;
}
/**
* Arima coefficients.
*/
export interface Schema$ArimaCoefficients {
/**
* Auto-regressive coefficients, an array of double.
*/
autoRegressiveCoefficients?: number[] | null;
/**
* Intercept coefficient, just a double not an array.
*/
interceptCoefficient?: number | null;
/**
* Moving-average coefficients, an array of double.
*/
movingAverageCoefficients?: number[] | null;
}
/**
* ARIMA model fitting metrics.
*/
export interface Schema$ArimaFittingMetrics {
/**
* AIC.
*/
aic?: number | null;
/**
* Log-likelihood.
*/
logLikelihood?: number | null;
/**
* Variance.
*/
variance?: number | null;
}
/**
* Model evaluation metrics for ARIMA forecasting models.
*/
export interface Schema$ArimaForecastingMetrics {
/**
* Arima model fitting metrics.
*/
arimaFittingMetrics?: Schema$ArimaFittingMetrics[];
/**
* Repeated as there can be many metric sets (one for each model) in auto-arima and the large-scale case.
*/
arimaSingleModelForecastingMetrics?: Schema$ArimaSingleModelForecastingMetrics[];
/**
* Whether Arima model fitted with drift or not. It is always false when d is not 1.
*/
hasDrift?: boolean[] | null;
/**
* Non-seasonal order.
*/
nonSeasonalOrder?: Schema$ArimaOrder[];
/**
* Seasonal periods. Repeated because multiple periods are supported for one time series.
*/
seasonalPeriods?: string[] | null;
/**
* Id to differentiate different time series for the large-scale case.
*/
timeSeriesId?: string[] | null;
}
/**
* Arima model information.
*/
export interface Schema$ArimaModelInfo {
/**
* Arima coefficients.
*/
arimaCoefficients?: Schema$ArimaCoefficients;
/**
* Arima fitting metrics.
*/
arimaFittingMetrics?: Schema$ArimaFittingMetrics;
/**
* Whether Arima model fitted with drift or not. It is always false when d is not 1.
*/
hasDrift?: boolean | null;
/**
* If true, holiday_effect is a part of time series decomposition result.
*/
hasHolidayEffect?: boolean | null;
/**
* If true, spikes_and_dips is a part of time series decomposition result.
*/
hasSpikesAndDips?: boolean | null;
/**
* If true, step_changes is a part of time series decomposition result.
*/
hasStepChanges?: boolean | null;
/**
* Non-seasonal order.
*/
nonSeasonalOrder?: Schema$ArimaOrder;
/**
* Seasonal periods. Repeated because multiple periods are supported for one time series.
*/
seasonalPeriods?: string[] | null;
/**
* The time_series_id value for this time series. It will be one of the unique values from the time_series_id_column specified during ARIMA model training. Only present when time_series_id_column training option was used.
*/
timeSeriesId?: string | null;
/**
* The tuple of time_series_ids identifying this time series. It will be one of the unique tuples of values present in the time_series_id_columns specified during ARIMA model training. Only present when time_series_id_columns training option was used and the order of values here are same as the order of time_series_id_columns.
*/
timeSeriesIds?: string[] | null;
}
/**
* Arima order, can be used for both non-seasonal and seasonal parts.
*/
export interface Schema$ArimaOrder {
/**
* Order of the differencing part.
*/
d?: string | null;
/**
* Order of the autoregressive part.
*/
p?: string | null;
/**
* Order of the moving-average part.
*/
q?: string | null;
}
/**
* (Auto-)arima fitting result. Wrap everything in ArimaResult for easier refactoring if we want to use model-specific iteration results.
*/
export interface Schema$ArimaResult {
/**
* This message is repeated because there are multiple arima models fitted in auto-arima. For non-auto-arima model, its size is one.
*/
arimaModelInfo?: Schema$ArimaModelInfo[];
/**
* Seasonal periods. Repeated because multiple periods are supported for one time series.
*/
seasonalPeriods?: string[] | null;
}
/**
* Model evaluation metrics for a single ARIMA forecasting model.
*/
export interface Schema$ArimaSingleModelForecastingMetrics {
/**
* Arima fitting metrics.
*/
arimaFittingMetrics?: Schema$ArimaFittingMetrics;
/**
* Is arima model fitted with drift or not. It is always false when d is not 1.
*/
hasDrift?: boolean | null;
/**
* If true, holiday_effect is a part of time series decomposition result.
*/
hasHolidayEffect?: boolean | null;
/**
* If true, spikes_and_dips is a part of time series decomposition result.
*/
hasSpikesAndDips?: boolean | null;
/**
* If true, step_changes is a part of time series decomposition result.
*/
hasStepChanges?: boolean | null;
/**
* Non-seasonal order.
*/
nonSeasonalOrder?: Schema$ArimaOrder;
/**
* Seasonal periods. Repeated because multiple periods are supported for one time series.
*/
seasonalPeriods?: string[] | null;
/**
* The time_series_id value for this time series. It will be one of the unique values from the time_series_id_column specified during ARIMA model training. Only present when time_series_id_column training option was used.
*/
timeSeriesId?: string | null;
/**
* The tuple of time_series_ids identifying this time series. It will be one of the unique tuples of values present in the time_series_id_columns specified during ARIMA model training. Only present when time_series_id_columns training option was used and the order of values here are same as the order of time_series_id_columns.
*/
timeSeriesIds?: string[] | null;
}
/**
* Specifies the audit configuration for a service. The configuration determines which permission types are logged, and what identities, if any, are exempted from logging. An AuditConfig must have one or more AuditLogConfigs. If there are AuditConfigs for both `allServices` and a specific service, the union of the two AuditConfigs is used for that service: the log_types specified in each AuditConfig are enabled, and the exempted_members in each AuditLogConfig are exempted. Example Policy with multiple AuditConfigs: { "audit_configs": [ { "service": "allServices", "audit_log_configs": [ { "log_type": "DATA_READ", "exempted_members": [ "user:jose@example.com" ] \}, { "log_type": "DATA_WRITE" \}, { "log_type": "ADMIN_READ" \} ] \}, { "service": "sampleservice.googleapis.com", "audit_log_configs": [ { "log_type": "DATA_READ" \}, { "log_type": "DATA_WRITE", "exempted_members": [ "user:aliya@example.com" ] \} ] \} ] \} For sampleservice, this policy enables DATA_READ, DATA_WRITE and ADMIN_READ logging. It also exempts `jose@example.com` from DATA_READ logging, and `aliya@example.com` from DATA_WRITE logging.
*/
export interface Schema$AuditConfig {
/**
* The configuration for logging of each type of permission.
*/
auditLogConfigs?: Schema$AuditLogConfig[];
/**
* Specifies a service that will be enabled for audit logging. For example, `storage.googleapis.com`, `cloudsql.googleapis.com`. `allServices` is a special value that covers all services.
*/
service?: string | null;
}
/**
* Provides the configuration for logging a type of permissions. Example: { "audit_log_configs": [ { "log_type": "DATA_READ", "exempted_members": [ "user:jose@example.com" ] \}, { "log_type": "DATA_WRITE" \} ] \} This enables 'DATA_READ' and 'DATA_WRITE' logging, while exempting jose@example.com from DATA_READ logging.
*/
export interface Schema$AuditLogConfig {
/**
* Specifies the identities that do not cause logging for this type of permission. Follows the same format of Binding.members.
*/
exemptedMembers?: string[] | null;
/**
* The log type that this config enables.
*/
logType?: string | null;
}
export interface Schema$AvroOptions {
/**
* [Optional] If sourceFormat is set to "AVRO", indicates whether to interpret logical types as the corresponding BigQuery data type (for example, TIMESTAMP), instead of using the raw type (for example, INTEGER).
*/
useAvroLogicalTypes?: boolean | null;
}
export interface Schema$BiEngineReason {
/**
* [Output-only] High-level BI Engine reason for partial or disabled acceleration.
*/
code?: string | null;
/**
* [Output-only] Free form human-readable reason for partial or disabled acceleration.
*/
message?: string | null;
}
export interface Schema$BiEngineStatistics {
/**
* [Output-only] Specifies which mode of BI Engine acceleration was performed (if any).
*/
accelerationMode?: string | null;
/**
* [Output-only] Specifies which mode of BI Engine acceleration was performed (if any).
*/
biEngineMode?: string | null;
/**
* In case of DISABLED or PARTIAL bi_engine_mode, these contain the explanatory reasons as to why BI Engine could not accelerate. In case the full query was accelerated, this field is not populated.
*/
biEngineReasons?: Schema$BiEngineReason[];
}
export interface Schema$BigQueryModelTraining {
/**
* [Output-only, Beta] Index of current ML training iteration. Updated during create model query job to show job progress.
*/
currentIteration?: number | null;
/**
* [Output-only, Beta] Expected number of iterations for the create model query job specified as num_iterations in the input query. The actual total number of iterations may be less than this number due to early stop.
*/
expectedTotalIterations?: string | null;
}
export interface Schema$BigtableColumn {
/**
* [Optional] The encoding of the values when the type is not STRING. Acceptable encoding values are: TEXT - indicates values are alphanumeric text strings. BINARY - indicates values are encoded using HBase Bytes.toBytes family of functions. 'encoding' can also be set at the column family level. However, the setting at this level takes precedence if 'encoding' is set at both levels.
*/
encoding?: string | null;
/**
* [Optional] If the qualifier is not a valid BigQuery field identifier i.e. does not match [a-zA-Z][a-zA-Z0-9_]*, a valid identifier must be provided as the column field name and is used as field name in queries.
*/
fieldName?: string | null;
/**
* [Optional] If this is set, only the latest version of value in this column are exposed. 'onlyReadLatest' can also be set at the column family level. However, the setting at this level takes precedence if 'onlyReadLatest' is set at both levels.
*/
onlyReadLatest?: boolean | null;
/**
* [Required] Qualifier of the column. Columns in the parent column family that has this exact qualifier are exposed as . field. If the qualifier is valid UTF-8 string, it can be specified in the qualifier_string field. Otherwise, a base-64 encoded value must be set to qualifier_encoded. The column field name is the same as the column qualifier. However, if the qualifier is not a valid BigQuery field identifier i.e. does not match [a-zA-Z][a-zA-Z0-9_]*, a valid identifier must be provided as field_name.
*/
qualifierEncoded?: string | null;
qualifierString?: string | null;
/**
* [Optional] The type to convert the value in cells of this column. The values are expected to be encoded using HBase Bytes.toBytes function when using the BINARY encoding value. Following BigQuery types are allowed (case-sensitive) - BYTES STRING INTEGER FLOAT BOOLEAN Default type is BYTES. 'type' can also be set at the column family level. However, the setting at this level takes precedence if 'type' is set at both levels.
*/
type?: string | null;
}
export interface Schema$BigtableColumnFamily {
/**
* [Optional] Lists of columns that should be exposed as individual fields as opposed to a list of (column name, value) pairs. All columns whose qualifier matches a qualifier in this list can be accessed as .. Other columns can be accessed as a list through .Column field.
*/
columns?: Schema$BigtableColumn[];
/**
* [Optional] The encoding of the values when the type is not STRING. Acceptable encoding values are: TEXT - indicates values are alphanumeric text strings. BINARY - indicates values are encoded using HBase Bytes.toBytes family of functions. This can be overridden for a specific column by listing that column in 'columns' and specifying an encoding for it.
*/
encoding?: string | null;
/**
* Identifier of the column family.
*/
familyId?: string | null;
/**
* [Optional] If this is set only the latest version of value are exposed for all columns in this column family. This can be overridden for a specific column by listing that column in 'columns' and specifying a different setting for that column.
*/
onlyReadLatest?: boolean | null;
/**
* [Optional] The type to convert the value in cells of this column family. The values are expected to be encoded using HBase Bytes.toBytes function when using the BINARY encoding value. Following BigQuery types are allowed (case-sensitive) - BYTES STRING INTEGER FLOAT BOOLEAN Default type is BYTES. This can be overridden for a specific column by listing that column in 'columns' and specifying a type for it.
*/
type?: string | null;
}
export interface Schema$BigtableOptions {
/**
* [Optional] List of column families to expose in the table schema along with their types. This list restricts the column families that can be referenced in queries and specifies their value types. You can use this list to do type conversions - see the 'type' field for more details. If you leave this list empty, all column families are present in the table schema and their values are read as BYTES. During a query only the column families referenced in that query are read from Bigtable.
*/
columnFamilies?: Schema$BigtableColumnFamily[];
/**
* [Optional] If field is true, then the column families that are not specified in columnFamilies list are not exposed in the table schema. Otherwise, they are read with BYTES type values. The default value is false.
*/
ignoreUnspecifiedColumnFamilies?: boolean | null;
/**
* [Optional] If field is true, then the rowkey column families will be read and converted to string. Otherwise they are read with BYTES type values and users need to manually cast them with CAST if necessary. The default value is false.
*/
readRowkeyAsString?: boolean | null;
}
/**
* Evaluation metrics for binary classification/classifier models.
*/
export interface Schema$BinaryClassificationMetrics {
/**
* Aggregate classification metrics.
*/
aggregateClassificationMetrics?: Schema$AggregateClassificationMetrics;
/**
* Binary confusion matrix at multiple thresholds.
*/
binaryConfusionMatrixList?: Schema$BinaryConfusionMatrix[];
/**
* Label representing the negative class.
*/
negativeLabel?: string | null;
/**
* Label representing the positive class.
*/
positiveLabel?: string | null;
}
/**
* Confusion matrix for binary classification models.
*/
export interface Schema$BinaryConfusionMatrix {
/**
* The fraction of predictions given the correct label.
*/
accuracy?: number | null;
/**
* The equally weighted average of recall and precision.
*/
f1Score?: number | null;
/**
* Number of false samples predicted as false.
*/
falseNegatives?: string | null;
/**
* Number of false samples predicted as true.
*/
falsePositives?: string | null;
/**
* Threshold value used when computing each of the following metric.
*/
positiveClassThreshold?: number | null;
/**
* The fraction of actual positive predictions that had positive actual labels.
*/
precision?: number | null;
/**
* The fraction of actual positive labels that were given a positive prediction.
*/
recall?: number | null;
/**
* Number of true samples predicted as false.
*/
trueNegatives?: string | null;
/**
* Number of true samples predicted as true.
*/
truePositives?: string | null;
}
/**
* Associates `members`, or principals, with a `role`.
*/
export interface Schema$Binding {
/**
* The condition that is associated with this binding. If the condition evaluates to `true`, then this binding applies to the current request. If the condition evaluates to `false`, then this binding does not apply to the current request. However, a different role binding might grant the same role to one or more of the principals in this binding. To learn which resources support conditions in their IAM policies, see the [IAM documentation](https://cloud.google.com/iam/help/conditions/resource-policies).
*/
condition?: Schema$Expr;
/**
* Specifies the principals requesting access for a Google Cloud resource. `members` can have the following values: * `allUsers`: A special identifier that represents anyone who is on the internet; with or without a Google account. * `allAuthenticatedUsers`: A special identifier that represents anyone who is authenticated with a Google account or a service account. Does not include identities that come from external identity providers (IdPs) through identity federation. * `user:{emailid\}`: An email address that represents a specific Google account. For example, `alice@example.com` . * `serviceAccount:{emailid\}`: An email address that represents a Google service account. For example, `my-other-app@appspot.gserviceaccount.com`. * `serviceAccount:{projectid\}.svc.id.goog[{namespace\}/{kubernetes-sa\}]`: An identifier for a [Kubernetes service account](https://cloud.google.com/kubernetes-engine/docs/how-to/kubernetes-service-accounts). For example, `my-project.svc.id.goog[my-namespace/my-kubernetes-sa]`. * `group:{emailid\}`: An email address that represents a Google group. For example, `admins@example.com`. * `domain:{domain\}`: The G Suite domain (primary) that represents all the users of that domain. For example, `google.com` or `example.com`. * `deleted:user:{emailid\}?uid={uniqueid\}`: An email address (plus unique identifier) representing a user that has been recently deleted. For example, `alice@example.com?uid=123456789012345678901`. If the user is recovered, this value reverts to `user:{emailid\}` and the recovered user retains the role in the binding. * `deleted:serviceAccount:{emailid\}?uid={uniqueid\}`: An email address (plus unique identifier) representing a service account that has been recently deleted. For example, `my-other-app@appspot.gserviceaccount.com?uid=123456789012345678901`. If the service account is undeleted, this value reverts to `serviceAccount:{emailid\}` and the undeleted service account retains the role in the binding. * `deleted:group:{emailid\}?uid={uniqueid\}`: An email address (plus unique identifier) representing a Google group that has been recently deleted. For example, `admins@example.com?uid=123456789012345678901`. If the group is recovered, this value reverts to `group:{emailid\}` and the recovered group retains the role in the binding.
*/
members?: string[] | null;
/**
* Role that is assigned to the list of `members`, or principals. For example, `roles/viewer`, `roles/editor`, or `roles/owner`.
*/
role?: string | null;
}
export interface Schema$BqmlIterationResult {
/**
* [Output-only, Beta] Time taken to run the training iteration in milliseconds.
*/
durationMs?: string | null;
/**
* [Output-only, Beta] Eval loss computed on the eval data at the end of the iteration. The eval loss is used for early stopping to avoid overfitting. No eval loss if eval_split_method option is specified as no_split or auto_split with input data size less than 500 rows.
*/
evalLoss?: number | null;
/**
* [Output-only, Beta] Index of the ML training iteration, starting from zero for each training run.
*/
index?: number | null;
/**
* [Output-only, Beta] Learning rate used for this iteration, it varies for different training iterations if learn_rate_strategy option is not constant.
*/
learnRate?: number | null;
/**
* [Output-only, Beta] Training loss computed on the training data at the end of the iteration. The training loss function is defined by model type.
*/
trainingLoss?: number | null;
}
export interface Schema$BqmlTrainingRun {
/**
* [Output-only, Beta] List of each iteration results.
*/
iterationResults?: Schema$BqmlIterationResult[];
/**
* [Output-only, Beta] Training run start time in milliseconds since the epoch.
*/
startTime?: string | null;
/**
* [Output-only, Beta] Different state applicable for a training run. IN PROGRESS: Training run is in progress. FAILED: Training run ended due to a non-retryable failure. SUCCEEDED: Training run successfully completed. CANCELLED: Training run cancelled by the user.
*/
state?: string | null;
/**
* [Output-only, Beta] Training options used by this training run. These options are mutable for subsequent training runs. Default values are explicitly stored for options not specified in the input query of the first training run. For subsequent training runs, any option not explicitly specified in the input query will be copied from the previous training run.
*/
trainingOptions?: {
earlyStop?: boolean;
l1Reg?: number;
l2Reg?: number;
learnRate?: number;
learnRateStrategy?: string;
lineSearchInitLearnRate?: number;
maxIteration?: string;
minRelProgress?: number;
warmStart?: boolean;
} | null;
}
/**
* Representative value of a categorical feature.
*/
export interface Schema$CategoricalValue {
/**
* Counts of all categories for the categorical feature. If there are more than ten categories, we return top ten (by count) and return one more CategoryCount with category "_OTHER_" and count as aggregate counts of remaining categories.
*/
categoryCounts?: Schema$CategoryCount[];
}
/**
* Represents the count of a single category within the cluster.
*/
export interface Schema$CategoryCount {
/**
* The name of category.
*/
category?: string | null;
/**
* The count of training samples matching the category within the cluster.
*/
count?: string | null;
}
export interface Schema$CloneDefinition {
/**
* [Required] Reference describing the ID of the table that was cloned.
*/
baseTableReference?: Schema$TableReference;
/**
* [Required] The time at which the base table was cloned. This value is reported in the JSON response using RFC3339 format.
*/
cloneTime?: string | null;
}
/**
* Message containing the information about one cluster.
*/
export interface Schema$Cluster {
/**
* Centroid id.
*/
centroidId?: string | null;
/**
* Count of training data rows that were assigned to this cluster.
*/
count?: string | null;
/**
* Values of highly variant features for this cluster.
*/
featureValues?: Schema$FeatureValue[];
}
/**
* Information about a single cluster for clustering model.
*/
export interface Schema$ClusterInfo {
/**
* Centroid id.
*/
centroidId?: string | null;
/**
* Cluster radius, the average distance from centroid to each point assigned to the cluster.
*/
clusterRadius?: number | null;
/**
* Cluster size, the total number of points assigned to the cluster.
*/
clusterSize?: string | null;
}
export interface Schema$Clustering {
/**
* [Repeated] One or more fields on which data should be clustered. Only top-level, non-repeated, simple-type fields are supported. When you cluster a table using multiple columns, the order of columns you specify is important. The order of the specified columns determines the sort order of the data.
*/
fields?: string[] | null;
}
/**
* Evaluation metrics for clustering models.
*/
export interface Schema$ClusteringMetrics {
/**
* Information for all clusters.
*/
clusters?: Schema$Cluster[];
/**
* Davies-Bouldin index.
*/
daviesBouldinIndex?: number | null;
/**
* Mean of squared distances between each sample to its cluster centroid.
*/
meanSquaredDistance?: number | null;
}
/**
* Confusion matrix for multi-class classification models.
*/
export interface Schema$ConfusionMatrix {
/**
* Confidence threshold used when computing the entries of the confusion matrix.
*/
confidenceThreshold?: number | null;
/**
* One row per actual label.
*/
rows?: Schema$Row[];
}
export interface Schema$ConnectionProperty {
/**
* [Required] Name of the connection property to set.
*/
key?: string | null;
/**
* [Required] Value of the connection property.
*/
value?: string | null;
}
export interface Schema$CsvOptions {
/**
* [Optional] Indicates if BigQuery should accept rows that are missing trailing optional columns. If true, BigQuery treats missing trailing columns as null values. If false, records with missing trailing columns are treated as bad records, and if there are too many bad records, an invalid error is returned in the job result. The default value is false.
*/
allowJaggedRows?: boolean | null;
/**
* [Optional] Indicates if BigQuery should allow quoted data sections that contain newline characters in a CSV file. The default value is false.
*/
allowQuotedNewlines?: boolean | null;
/**
* [Optional] The character encoding of the data. The supported values are UTF-8 or ISO-8859-1. The default value is UTF-8. BigQuery decodes the data after the raw, binary data has been split using the values of the quote and fieldDelimiter properties.
*/
encoding?: string | null;
/**
* [Optional] The separator for fields in a CSV file. BigQuery converts the string to ISO-8859-1 encoding, and then uses the first byte of the encoded string to split the data in its raw, binary state. BigQuery also supports the escape sequence "\t" to specify a tab separator. The default value is a comma (',').
*/
fieldDelimiter?: string | null;
/**
* [Optional] An custom string that will represent a NULL value in CSV import data.
*/
null_marker?: string | null;
/**
* [Optional] Preserves the embedded ASCII control characters (the first 32 characters in the ASCII-table, from '\x00' to '\x1F') when loading from CSV. Only applicable to CSV, ignored for other formats.
*/
preserveAsciiControlCharacters?: boolean | null;
/**
* [Optional] The value that is used to quote data sections in a CSV file. BigQuery converts the string to ISO-8859-1 encoding, and then uses the first byte of the encoded string to split the data in its raw, binary state. The default value is a double-quote ('"'). If your data does not contain quoted sections, set the property value to an empty string. If your data contains quoted newline characters, you must also set the allowQuotedNewlines property to true.
*/
quote?: string | null;
/**
* [Optional] The number of rows at the top of a CSV file that BigQuery will skip when reading the data. The default value is 0. This property is useful if you have header rows in the file that should be skipped. When autodetect is on, the behavior is the following: * skipLeadingRows unspecified - Autodetect tries to detect headers in the first row. If they are not detected, the row is read as data. Otherwise data is read starting from the second row. * skipLeadingRows is 0 - Instructs autodetect that there are no headers and data should be read starting from the first row. * skipLeadingRows = N \> 0 - Autodetect skips N-1 rows and tries to detect headers in row N. If headers are not detected, row N is just skipped. Otherwise row N is used to extract column names for the detected schema.
*/
skipLeadingRows?: string | null;
}
export interface Schema$DataMaskingStatistics {
/**
* [Output-only] [Preview] Whether any accessed data was protected by data masking. The actual evaluation is done by accessStats.masked_field_count \> 0. Since this is only used for the discovery_doc generation purpose, as long as the type (boolean) matches, client library can leverage this. The actual evaluation of the variable is done else-where.
*/
dataMaskingApplied?: boolean | null;
}
export interface Schema$Dataset {
/**
* [Optional] An array of objects that define dataset access for one or more entities. You can set this property when inserting or updating a dataset in order to control who is allowed to access the data. If unspecified at dataset creation time, BigQuery adds default dataset access for the following entities: access.specialGroup: projectReaders; access.role: READER; access.specialGroup: projectWriters; access.role: WRITER; access.specialGroup: projectOwners; access.role: OWNER; access.userByEmail: [dataset creator email]; access.role: OWNER;
*/
access?: Array<{
dataset?: Schema$DatasetAccessEntry;
domain?: string;
groupByEmail?: string;
iamMember?: string;
role?: string;
routine?: Schema$RoutineReference;
specialGroup?: string;
userByEmail?: string;
view?: Schema$TableReference;
}> | null;
/**
* [Output-only] The time when this dataset was created, in milliseconds since the epoch.
*/
creationTime?: string | null;
/**
* [Required] A reference that identifies the dataset.
*/
datasetReference?: Schema$DatasetReference;
/**
* [Output-only] The default collation of the dataset.
*/
defaultCollation?: string | null;
defaultEncryptionConfiguration?: Schema$EncryptionConfiguration;
/**
* [Optional] The default partition expiration for all partitioned tables in the dataset, in milliseconds. Once this property is set, all newly-created partitioned tables in the dataset will have an expirationMs property in the timePartitioning settings set to this value, and changing the value will only affect new tables, not existing ones. The storage in a partition will have an expiration time of its partition time plus this value. Setting this property overrides the use of defaultTableExpirationMs for partitioned tables: only one of defaultTableExpirationMs and defaultPartitionExpirationMs will be used for any new partitioned table. If you provide an explicit timePartitioning.expirationMs when creating or updating a partitioned table, that value takes precedence over the default partition expiration time indicated by this property.
*/
defaultPartitionExpirationMs?: string | null;
/**
* [Output-only] The default rounding mode of the dataset.
*/
defaultRoundingMode?: string | null;
/**
* [Optional] The default lifetime of all tables in the dataset, in milliseconds. The minimum value is 3600000 milliseconds (one hour). Once this property is set, all newly-created tables in the dataset will have an expirationTime property set to the creation time plus the value in this property, and changing the value will only affect new tables, not existing ones. When the expirationTime for a given table is reached, that table will be deleted automatically. If a table's expirationTime is modified or removed before the table expires, or if you provide an explicit expirationTime when creating a table, that value takes precedence over the default expiration time indicated by this property.
*/
defaultTableExpirationMs?: string | null;
/**
* [Optional] A user-friendly description of the dataset.
*/
description?: string | null;
/**
* [Output-only] A hash of the resource.
*/
etag?: string | null;
/**
* [Optional] A descriptive name for the dataset.
*/
friendlyName?: string | null;
/**
* [Output-only] The fully-qualified unique name of the dataset in the format projectId:datasetId. The dataset name without the project name is given in the datasetId field. When creating a new dataset, leave this field blank, and instead specify the datasetId field.
*/
id?: string | null;
/**
* [Optional] Indicates if table names are case insensitive in the dataset.
*/
isCaseInsensitive?: boolean | null;
/**
* [Output-only] The resource type.
*/
kind?: string | null;
/**
* The labels associated with this dataset. You can use these to organize and group your datasets. You can set this property when inserting or updating a dataset. See Creating and Updating Dataset Labels for more information.
*/
labels?: {
[key: string]: string;
} | null;
/**
* [Output-only] The date when this dataset or any of its tables was last modified, in milliseconds since the epoch.
*/
lastModifiedTime?: string | null;
/**
* The geographic location where the dataset should reside. The default value is US. See details at https://cloud.google.com/bigquery/docs/locations.
*/
location?: string | null;
/**
* [Optional] Number of hours for the max time travel for all tables in the dataset.
*/
maxTimeTravelHours?: string | null;
/**
* [Output-only] Reserved for future use.
*/
satisfiesPzs?: boolean | null;
/**
* [Output-only] A URL that can be used to access the resource again. You can use this URL in Get or Update requests to the resource.
*/
selfLink?: string | null;
/**
* [Optional] Storage billing model to be used for all tables in the dataset. Can be set to PHYSICAL. Default is LOGICAL.
*/
storageBillingModel?: string | null;
/**
* [Optional]The tags associated with this dataset. Tag keys are globally unique.
*/
tags?: Array<{
tagKey?: string;
tagValue?: string;
}> | null;
}
export interface Schema$DatasetAccessEntry {
/**
* [Required] The dataset this entry applies to.
*/
dataset?: Schema$DatasetReference;
targetTypes?: string[] | null;
}
export interface Schema$DatasetList {
/**
* An array of the dataset resources in the project. Each resource contains basic information. For full information about a particular dataset resource, use the Datasets: get method. This property is omitted when there are no datasets in the project.
*/
datasets?: Array<{
datasetReference?: Schema$DatasetReference;
friendlyName?: string;
id?: string;
kind?: string;
labels?: {
[key: string]: string;
};
location?: string;
}> | null;
/**
* A hash value of the results page. You can use this property to determine if the page has changed since the last request.
*/
etag?: string | null;
/**
* The list type. This property always returns the value "bigquery#datasetList".
*/
kind?: string | null;
/**
* A token that can be used to request the next results page. This property is omitted on the final results page.
*/
nextPageToken?: string | null;
}
export interface Schema$DatasetReference {
/**
* [Required] A unique ID for this dataset, without the project name. The ID must contain only letters (a-z, A-Z), numbers (0-9), or underscores (_). The maximum length is 1,024 characters.
*/
datasetId?: string | null;
/**
* [Optional] The ID of the project containing this dataset.
*/
projectId?: string | null;
}
/**
* Data split result. This contains references to the training and evaluation data tables that were used to train the model.
*/
export interface Schema$DataSplitResult {
/**
* Table reference of the evaluation data after split.
*/
evaluationTable?: Schema$TableReference;
/**
* Table reference of the test data after split.
*/
testTable?: Schema$TableReference;
/**
* Table reference of the training data after split.
*/
trainingTable?: Schema$TableReference;
}
export interface Schema$DestinationTableProperties {
/**
* [Optional] The description for the destination table. This will only be used if the destination table is newly created. If the table already exists and a value different than the current description is provided, the job will fail.
*/
description?: string | null;
/**
* [Internal] This field is for Google internal use only.
*/
expirationTime?: string | null;
/**
* [Optional] The friendly name for the destination table. This will only be used if the destination table is newly created. If the table already exists and a value different than the current friendly name is provided, the job will fail.
*/
friendlyName?: string | null;
/**
* [Optional] The labels associated with this table. You can use these to organize and group your tables. This will only be used if the destination table is newly created. If the table already exists and labels are different than the current labels are provided, the job will fail.
*/
labels?: {
[key: string]: string;
} | null;
}
/**
* Model evaluation metrics for dimensionality reduction models.
*/
export interface Schema$DimensionalityReductionMetrics {
/**
* Total percentage of variance explained by the selected principal components.
*/
totalExplainedVarianceRatio?: number | null;
}
export interface Schema$DmlStatistics {
/**
* Number of deleted Rows. populated by DML DELETE, MERGE and TRUNCATE statements.
*/
deletedRowCount?: string | null;
/**
* Number of inserted Rows. Populated by DML INSERT and MERGE statements.
*/
insertedRowCount?: string | null;
/**
* Number of updated Rows. Populated by DML UPDATE and MERGE statements.
*/
updatedRowCount?: string | null;
}
/**
* Discrete candidates of a double hyperparameter.
*/
export interface Schema$DoubleCandidates {
/**
* Candidates for the double parameter in increasing order.
*/
candidates?: number[] | null;
}
/**
* Search space for a double hyperparameter.
*/
export interface Schema$DoubleHparamSearchSpace {
/**
* Candidates of the double hyperparameter.
*/
candidates?: Schema$DoubleCandidates;
/**
* Range of the double hyperparameter.
*/
range?: Schema$DoubleRange;
}
/**
* Range of a double hyperparameter.
*/
export interface Schema$DoubleRange {
/**
* Max value of the double parameter.
*/
max?: number | null;
/**
* Min value of the double parameter.
*/
min?: number | null;
}
export interface Schema$EncryptionConfiguration {
/**
* Optional. Describes the Cloud KMS encryption key that will be used to protect destination BigQuery table. The BigQuery Service Account associated with your project requires access to this encryption key.
*/
kmsKeyName?: string | null;
}
/**
* A single entry in the confusion matrix.
*/
export interface Schema$Entry {
/**
* Number of items being predicted as this label.
*/
itemCount?: string | null;
/**
* The predicted label. For confidence_threshold \> 0, we will also add an entry indicating the number of items under the confidence threshold.
*/
predictedLabel?: string | null;
}
export interface Schema$ErrorProto {
/**
* Debugging information. This property is internal to Google and should not be used.
*/
debugInfo?: string | null;
/**
* Specifies where the error occurred, if present.
*/
location?: string | null;
/**
* A human-readable description of the error.
*/
message?: string | null;
/**
* A short error code that summarizes the error.
*/
reason?: string | null;
}
/**
* Evaluation metrics of a model. These are either computed on all training data or just the eval data based on whether eval data was used during training. These are not present for imported models.
*/
export interface Schema$EvaluationMetrics {
/**
* Populated for ARIMA models.
*/
arimaForecastingMetrics?: Schema$ArimaForecastingMetrics;
/**
* Populated for binary classification/classifier models.
*/
binaryClassificationMetrics?: Schema$BinaryClassificationMetrics;
/**
* Populated for clustering models.
*/
clusteringMetrics?: Schema$ClusteringMetrics;
/**
* Evaluation metrics when the model is a dimensionality reduction model, which currently includes PCA.
*/
dimensionalityReductionMetrics?: Schema$DimensionalityReductionMetrics;
/**
* Populated for multi-class classification/classifier models.
*/
multiClassClassificationMetrics?: Schema$MultiClassClassificationMetrics;
/**
* Populated for implicit feedback type matrix factorization models.
*/
rankingMetrics?: Schema$RankingMetrics;
/**
* Populated for regression models and explicit feedback type matrix factorization models.
*/
regressionMetrics?: Schema$RegressionMetrics;
}
export interface Schema$ExplainQueryStage {
/**