UNPKG

@firebase/firestore

Version:

The Cloud Firestore component of the Firebase JS SDK.

510 lines (509 loc) • 19 kB
/** * @license * Copyright 2017 Google LLC * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ import { BatchId, ListenSequenceNumber, TargetId } from '../core/types'; import { IndexKind } from '../model/field_index'; import { BundledQuery } from '../protos/firestore_bundle_proto'; import { Document as ProtoDocument, DocumentsTarget as ProtoDocumentsTarget, QueryTarget as ProtoQueryTarget, Write as ProtoWrite } from '../protos/firestore_proto_api'; import { EncodedResourcePath } from './encoded_resource_path'; import { DbTimestampKey, KeySafeBytes } from './indexeddb_sentinels'; /** * Schema Version for the Web client: * 1. Initial version including Mutation Queue, Query Cache, and Remote * Document Cache * 2. Used to ensure a targetGlobal object exists and add targetCount to it. No * longer required because migration 3 unconditionally clears it. * 3. Dropped and re-created Query Cache to deal with cache corruption related * to limbo resolution. Addresses * https://github.com/firebase/firebase-ios-sdk/issues/1548 * 4. Multi-Tab Support. * 5. Removal of held write acks. * 6. Create document global for tracking document cache size. * 7. Ensure every cached document has a sentinel row with a sequence number. * 8. Add collection-parent index for Collection Group queries. * 9. Change RemoteDocumentChanges store to be keyed by readTime rather than * an auto-incrementing ID. This is required for Index-Free queries. * 10. Rewrite the canonical IDs to the explicit Protobuf-based format. * 11. Add bundles and named_queries for bundle support. * 12. Add document overlays. * 13. Rewrite the keys of the remote document cache to allow for efficient * document lookup via `getAll()`. * 14. Add overlays. * 15. Add indexing support. * 16. Parse timestamp strings before creating index entries. * 17. TODO(tomandersen) * 18. Encode key safe representations of IndexEntry in DbIndexEntryStore. */ export declare const SCHEMA_VERSION = 18; /** * Wrapper class to store timestamps (seconds and nanos) in IndexedDb objects. */ export interface DbTimestamp { seconds: number; nanoseconds: number; } /** * A singleton object to be stored in the 'owner' store in IndexedDb. * * A given database can have a single primary tab assigned at a given time. That * tab must validate that it is still holding the primary lease before every * operation that requires locked access. The primary tab should regularly * write an updated timestamp to this lease to prevent other tabs from * "stealing" the primary lease */ export interface DbPrimaryClient { ownerId: string; /** Whether to allow shared access from multiple tabs. */ allowTabSynchronization: boolean; leaseTimestampMs: number; } /** * An object to be stored in the 'mutationQueues' store in IndexedDb. * * Each user gets a single queue of MutationBatches to apply to the server. * DbMutationQueue tracks the metadata about the queue. */ export interface DbMutationQueue { /** * The normalized user ID to which this queue belongs. */ userId: string; /** * An identifier for the highest numbered batch that has been acknowledged * by the server. All MutationBatches in this queue with batchIds less * than or equal to this value are considered to have been acknowledged by * the server. * * NOTE: this is deprecated and no longer used by the code. */ lastAcknowledgedBatchId: number; /** * A stream token that was previously sent by the server. * * See StreamingWriteRequest in datastore.proto for more details about * usage. * * After sending this token, earlier tokens may not be used anymore so * only a single stream token is retained. * * NOTE: this is deprecated and no longer used by the code. */ lastStreamToken: string; } /** * An object to be stored in the 'mutations' store in IndexedDb. * * Represents a batch of user-level mutations intended to be sent to the server * in a single write. Each user-level batch gets a separate DbMutationBatch * with a new batchId. */ export interface DbMutationBatch { /** * The normalized user ID to which this batch belongs. */ userId: string; /** * An identifier for this batch, allocated using an auto-generated key. */ batchId: BatchId; /** * The local write time of the batch, stored as milliseconds since the * epoch. */ localWriteTimeMs: number; /** * A list of "mutations" that represent a partial base state from when this * write batch was initially created. During local application of the write * batch, these baseMutations are applied prior to the real writes in order * to override certain document fields from the remote document cache. This * is necessary in the case of non-idempotent writes (e.g. `increment()` * transforms) to make sure that the local view of the modified documents * doesn't flicker if the remote document cache receives the result of the * non-idempotent write before the write is removed from the queue. * * These mutations are never sent to the backend. */ baseMutations?: ProtoWrite[]; /** * A list of mutations to apply. All mutations will be applied atomically. * * Mutations are serialized via toMutation(). */ mutations: ProtoWrite[]; } /** * An object to be stored in the 'documentMutations' store in IndexedDb. * * A manually maintained index of all the mutation batches that affect a given * document key. The rows in this table are references based on the contents of * DbMutationBatch.mutations. */ export interface DbDocumentMutation { } /** * Represents the known absence of a document at a particular version. * Stored in IndexedDb as part of a DbRemoteDocument object. */ export interface DbNoDocument { path: string[]; readTime: DbTimestamp; } /** * Represents a document that is known to exist but whose data is unknown. * Stored in IndexedDb as part of a DbRemoteDocument object. */ export interface DbUnknownDocument { path: string[]; version: DbTimestamp; } /** * An object to be stored in the 'remoteDocuments' store in IndexedDb. * It represents either: * * - A complete document. * - A "no document" representing a document that is known not to exist (at * some version). * - An "unknown document" representing a document that is known to exist (at * some version) but whose contents are unknown. * * The document key is split up across `prefixPath`, `collectionGroup` and * `documentId`. * * Note: This is the persisted equivalent of a MaybeDocument and could perhaps * be made more general if necessary. */ export interface DbRemoteDocument { /** The path to the document's collection (excluding). */ prefixPath: string[]; /** The collection ID the document is directly nested under. */ collectionGroup: string; /** The document ID. */ documentId: string; /** When the document was read from the backend. */ readTime: DbTimestampKey; /** * Set to an instance of DbUnknownDocument if the data for a document is * not known, but it is known that a document exists at the specified * version (e.g. it had a successful update applied to it) */ unknownDocument?: DbUnknownDocument; /** * Set to an instance of a DbNoDocument if it is known that no document * exists. */ noDocument?: DbNoDocument; /** * Set to an instance of a Document if there's a cached version of the * document. */ document?: ProtoDocument; /** * Documents that were written to the remote document store based on * a write acknowledgment are marked with `hasCommittedMutations`. These * documents are potentially inconsistent with the backend's copy and use * the write's commit version as their document version. */ hasCommittedMutations: boolean; } /** * Contains a single entry that has metadata about the remote document cache. */ export interface DbRemoteDocumentGlobal { /** * Approximately the total size in bytes of all the * documents in the document cache. */ byteSize: number; } /** * The persisted type for a query nested with in the 'targets' store in * IndexedDb. We use the proto definitions for these two kinds of queries in * order to avoid writing extra serialization logic. */ export type DbQuery = ProtoQueryTarget | ProtoDocumentsTarget; /** * An object to be stored in the 'targets' store in IndexedDb. * * This is based on and should be kept in sync with the proto used in the iOS * client. * * Each query the client listens to against the server is tracked on disk so * that the query can be efficiently resumed on restart. */ export interface DbTarget { /** * An auto-generated sequential numeric identifier for the query. * * Queries are stored using their canonicalId as the key, but these * canonicalIds can be quite long so we additionally assign a unique * queryId which can be used by referenced data structures (e.g. * indexes) to minimize the on-disk cost. */ targetId: TargetId; /** * The canonical string representing this query. This is not unique. */ canonicalId: string; /** * The last readTime received from the Watch Service for this query. * * This is the same value as TargetChange.read_time in the protos. */ readTime: DbTimestamp; /** * An opaque, server-assigned token that allows watching a query to be * resumed after disconnecting without retransmitting all the data * that matches the query. The resume token essentially identifies a * point in time from which the server should resume sending results. * * This is related to the snapshotVersion in that the resumeToken * effectively also encodes that value, but the resumeToken is opaque * and sometimes encodes additional information. * * A consequence of this is that the resumeToken should be used when * asking the server to reason about where this client is in the watch * stream, but the client should use the snapshotVersion for its own * purposes. * * This is the same value as TargetChange.resume_token in the protos. */ resumeToken: string; /** * A sequence number representing the last time this query was * listened to, used for garbage collection purposes. * * Conventionally this would be a timestamp value, but device-local * clocks are unreliable and they must be able to create new listens * even while disconnected. Instead this should be a monotonically * increasing number that's incremented on each listen call. * * This is different from the queryId since the queryId is an * immutable identifier assigned to the Query on first use while * lastListenSequenceNumber is updated every time the query is * listened to. */ lastListenSequenceNumber: number; /** * Denotes the maximum snapshot version at which the associated query view * contained no limbo documents. Undefined for data written prior to * schema version 9. */ lastLimboFreeSnapshotVersion?: DbTimestamp; /** * The query for this target. * * Because canonical ids are not unique we must store the actual query. We * use the proto to have an object we can persist without having to * duplicate translation logic to and from a `Query` object. */ query: DbQuery; } /** * An object representing an association between a target and a document, or a * sentinel row marking the last sequence number at which a document was used. * Each document cached must have a corresponding sentinel row before lru * garbage collection is enabled. * * The target associations and sentinel rows are co-located so that orphaned * documents and their sequence numbers can be identified efficiently via a scan * of this store. */ export interface DbTargetDocument { /** * The targetId identifying a target or 0 for a sentinel row. */ targetId: TargetId; /** * The path to the document, as encoded in the key. */ path: EncodedResourcePath; /** * If this is a sentinel row, this should be the sequence number of the last * time the document specified by `path` was used. Otherwise, it should be * `undefined`. */ sequenceNumber?: ListenSequenceNumber; } /** * A record of global state tracked across all Targets, tracked separately * to avoid the need for extra indexes. * * This should be kept in-sync with the proto used in the iOS client. */ export interface DbTargetGlobal { /** * The highest numbered target id across all targets. * * See DbTarget.targetId. */ highestTargetId: TargetId; /** * The highest numbered lastListenSequenceNumber across all targets. * * See DbTarget.lastListenSequenceNumber. */ highestListenSequenceNumber: number; /** * A global snapshot version representing the last consistent snapshot we * received from the backend. This is monotonically increasing and any * snapshots received from the backend prior to this version (e.g. for * targets resumed with a resumeToken) should be suppressed (buffered) * until the backend has caught up to this snapshot version again. This * prevents our cache from ever going backwards in time. */ lastRemoteSnapshotVersion: DbTimestamp; /** * The number of targets persisted. */ targetCount: number; } /** * An object representing an association between a Collection id (e.g. 'messages') * to a parent path (e.g. '/chats/123') that contains it as a (sub)collection. * This is used to efficiently find all collections to query when performing * a Collection Group query. */ export interface DbCollectionParent { /** * The collectionId (e.g. 'messages') */ collectionId: string; /** * The path to the parent (either a document location or an empty path for * a root-level collection). */ parent: EncodedResourcePath; } /** * A record of the metadata state of each client. * * PORTING NOTE: This is used to synchronize multi-tab state and does not need * to be ported to iOS or Android. */ export interface DbClientMetadata { /** The auto-generated client id assigned at client startup. */ clientId: string; /** The last time this state was updated. */ updateTimeMs: number; /** Whether the client's network connection is enabled. */ networkEnabled: boolean; /** Whether this client is running in a foreground tab. */ inForeground: boolean; } /** An object representing a bundle loaded by the SDK. */ export interface DbBundle { /** The ID of the loaded bundle. */ bundleId: string; /** The create time of the loaded bundle. */ createTime: DbTimestamp; /** The schema version of the loaded bundle. */ version: number; } /** An object representing a named query loaded by the SDK via a bundle. */ export interface DbNamedQuery { /** The name of the query. */ name: string; /** The read time of the results saved in the bundle from the named query. */ readTime: DbTimestamp; /** The query saved in the bundle. */ bundledQuery: BundledQuery; } /** An object representing the global configuration for a field index. */ export interface DbIndexConfiguration { /** * The index id for this entry. Undefined for indexes that are not yet * persisted. */ indexId?: number; /** The collection group this index belongs to. */ collectionGroup: string; /** The fields to index for this index. */ fields: Array<[name: string, kind: IndexKind]>; } /** * An object describing how up-to-date the index backfill is for each user and * index. */ export interface DbIndexState { /** The index id for this entry. */ indexId: number; /** The user id for this entry. */ uid: string; /** * A number that indicates when the index was last updated (relative to * other indexes). */ sequenceNumber: number; /** * The latest read time that has been indexed by Firestore for this field * index. Set to `{seconds: 0, nanos: 0}` if no documents have been indexed. */ readTime: DbTimestamp; /** * The last document that has been indexed for this field index. Empty if * no documents have been indexed. */ documentKey: EncodedResourcePath; /** * The largest mutation batch id that has been processed for this index. -1 * if no mutations have been indexed. */ largestBatchId: number; } /** An object that stores the encoded entries for all documents and fields. */ export interface DbIndexEntry { /** The index id for this entry. */ indexId: number; /** The user id for this entry. */ uid: string; /** The encoded array index value for this entry. */ arrayValue: KeySafeBytes; /** The encoded directional value for equality and inequality filters. */ directionalValue: KeySafeBytes; /** * The document key this entry points to. This entry is encoded by an ordered * encoder to match the key order of the index. */ orderedDocumentKey: KeySafeBytes; /** The segments of the document key this entry points to. */ documentKey: string[]; } /** * An object representing a document overlay. */ export interface DbDocumentOverlay { /** The user ID to whom this overlay belongs. */ userId: string; /** The path to the collection that contains the document. */ collectionPath: string; /** The ID (key) of the document within the collection. */ documentId: string; /** The collection group to which the document belongs. */ collectionGroup: string; /** The largest batch ID that's been applied for this overlay. */ largestBatchId: number; /** The overlay mutation. */ overlayMutation: ProtoWrite; } /** * An object containing global name/value pair. */ export interface DbGlobals { /** Name is a globally unique identifier for a value. */ name: string; /** Value is a general purpose storage for global data. */ value: Uint8Array; }