| 123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527 |
- /*
- * Copyright 2017 Google
- *
- * Licensed under the Apache License, Version 2.0 (the "License");
- * you may not use this file except in compliance with the License.
- * You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
- #import "Firestore/Source/Local/FSTLocalStore.h"
- #include <memory>
- #include <set>
- #include <unordered_map>
- #include <utility>
- #include <vector>
- #import "FIRTimestamp.h"
- #import "Firestore/Source/Core/FSTQuery.h"
- #import "Firestore/Source/Local/FSTLRUGarbageCollector.h"
- #import "Firestore/Source/Local/FSTPersistence.h"
- #import "Firestore/Source/Local/FSTQueryData.h"
- #import "Firestore/Source/Model/FSTDocument.h"
- #import "Firestore/Source/Model/FSTMutation.h"
- #import "Firestore/Source/Model/FSTMutationBatch.h"
- #include "Firestore/core/include/firebase/firestore/timestamp.h"
- #include "Firestore/core/src/firebase/firestore/auth/user.h"
- #include "Firestore/core/src/firebase/firestore/core/target_id_generator.h"
- #include "Firestore/core/src/firebase/firestore/immutable/sorted_set.h"
- #include "Firestore/core/src/firebase/firestore/local/local_documents_view.h"
- #include "Firestore/core/src/firebase/firestore/local/local_view_changes.h"
- #include "Firestore/core/src/firebase/firestore/local/local_write_result.h"
- #include "Firestore/core/src/firebase/firestore/local/mutation_queue.h"
- #include "Firestore/core/src/firebase/firestore/local/query_cache.h"
- #include "Firestore/core/src/firebase/firestore/local/reference_set.h"
- #include "Firestore/core/src/firebase/firestore/local/remote_document_cache.h"
- #include "Firestore/core/src/firebase/firestore/model/document_key_set.h"
- #include "Firestore/core/src/firebase/firestore/model/document_map.h"
- #include "Firestore/core/src/firebase/firestore/model/snapshot_version.h"
- #include "Firestore/core/src/firebase/firestore/remote/remote_event.h"
- #include "Firestore/core/src/firebase/firestore/util/hard_assert.h"
- #include "Firestore/core/src/firebase/firestore/util/log.h"
- #include "absl/memory/memory.h"
- #include "absl/types/optional.h"
- using firebase::Timestamp;
- using firebase::firestore::auth::User;
- using firebase::firestore::core::TargetIdGenerator;
- using firebase::firestore::local::LocalDocumentsView;
- using firebase::firestore::local::LocalViewChanges;
- using firebase::firestore::local::LocalWriteResult;
- using firebase::firestore::local::LruResults;
- using firebase::firestore::local::MutationQueue;
- using firebase::firestore::local::QueryCache;
- using firebase::firestore::local::ReferenceSet;
- using firebase::firestore::local::RemoteDocumentCache;
- using firebase::firestore::model::BatchId;
- using firebase::firestore::model::DocumentKey;
- using firebase::firestore::model::DocumentKeySet;
- using firebase::firestore::model::DocumentMap;
- using firebase::firestore::model::DocumentVersionMap;
- using firebase::firestore::model::FieldMask;
- using firebase::firestore::model::FieldPath;
- using firebase::firestore::model::ListenSequenceNumber;
- using firebase::firestore::model::MaybeDocumentMap;
- using firebase::firestore::model::ObjectValue;
- using firebase::firestore::model::Precondition;
- using firebase::firestore::model::SnapshotVersion;
- using firebase::firestore::model::TargetId;
- using firebase::firestore::remote::RemoteEvent;
- using firebase::firestore::remote::TargetChange;
- NS_ASSUME_NONNULL_BEGIN
- /**
- * The maximum time to leave a resume token buffered without writing it out. This value is
- * arbitrary: it's long enough to avoid several writes (possibly indefinitely if updates come more
- * frequently than this) but short enough that restarting after crashing will still have a pretty
- * recent resume token.
- */
- static const int64_t kResumeTokenMaxAgeSeconds = 5 * 60; // 5 minutes
- @interface FSTLocalStore ()
- /** Manages our in-memory or durable persistence. */
- @property(nonatomic, strong, readonly) id<FSTPersistence> persistence;
- /** Maps a query to the data about that query. */
- @property(nonatomic) QueryCache *queryCache;
- @end
- @implementation FSTLocalStore {
- /** Used to generate targetIDs for queries tracked locally. */
- TargetIdGenerator _targetIDGenerator;
- /** The set of all cached remote documents. */
- RemoteDocumentCache *_remoteDocumentCache;
- QueryCache *_queryCache;
- /** The set of all mutations that have been sent but not yet been applied to the backend. */
- MutationQueue *_mutationQueue;
- /** The "local" view of all documents (layering mutationQueue on top of remoteDocumentCache). */
- std::unique_ptr<LocalDocumentsView> _localDocuments;
- /** The set of document references maintained by any local views. */
- ReferenceSet _localViewReferences;
- /** Maps a targetID to data about its query. */
- std::unordered_map<TargetId, FSTQueryData *> _targetIDs;
- }
- - (instancetype)initWithPersistence:(id<FSTPersistence>)persistence
- initialUser:(const User &)initialUser {
- if (self = [super init]) {
- _persistence = persistence;
- _mutationQueue = [persistence mutationQueueForUser:initialUser];
- _remoteDocumentCache = [persistence remoteDocumentCache];
- _queryCache = [persistence queryCache];
- _localDocuments = absl::make_unique<LocalDocumentsView>(_remoteDocumentCache, _mutationQueue,
- [_persistence indexManager]);
- [_persistence.referenceDelegate addInMemoryPins:&_localViewReferences];
- _targetIDGenerator = TargetIdGenerator::QueryCacheTargetIdGenerator(0);
- }
- return self;
- }
- - (void)start {
- [self startMutationQueue];
- TargetId targetID = _queryCache->highest_target_id();
- _targetIDGenerator = TargetIdGenerator::QueryCacheTargetIdGenerator(targetID);
- }
- - (void)startMutationQueue {
- self.persistence.run("Start MutationQueue", [&]() { _mutationQueue->Start(); });
- }
- - (MaybeDocumentMap)userDidChange:(const User &)user {
- // Swap out the mutation queue, grabbing the pending mutation batches before and after.
- std::vector<FSTMutationBatch *> oldBatches = self.persistence.run(
- "OldBatches",
- [&]() -> std::vector<FSTMutationBatch *> { return _mutationQueue->AllMutationBatches(); });
- // The old one has a reference to the mutation queue, so nil it out first.
- _localDocuments.reset();
- _mutationQueue = [self.persistence mutationQueueForUser:user];
- [self startMutationQueue];
- return self.persistence.run("NewBatches", [&]() -> MaybeDocumentMap {
- std::vector<FSTMutationBatch *> newBatches = _mutationQueue->AllMutationBatches();
- // Recreate our LocalDocumentsView using the new MutationQueue.
- _localDocuments = absl::make_unique<LocalDocumentsView>(_remoteDocumentCache, _mutationQueue,
- [_persistence indexManager]);
- // Union the old/new changed keys.
- DocumentKeySet changedKeys;
- for (const std::vector<FSTMutationBatch *> &batches : {oldBatches, newBatches}) {
- for (FSTMutationBatch *batch : batches) {
- for (FSTMutation *mutation : [batch mutations]) {
- changedKeys = changedKeys.insert(mutation.key);
- }
- }
- }
- // Return the set of all (potentially) changed documents as the result of the user change.
- return _localDocuments->GetDocuments(changedKeys);
- });
- }
- - (LocalWriteResult)locallyWriteMutations:(std::vector<FSTMutation *> &&)mutations {
- Timestamp localWriteTime = Timestamp::Now();
- DocumentKeySet keys;
- for (FSTMutation *mutation : mutations) {
- keys = keys.insert(mutation.key);
- }
- return self.persistence.run("Locally write mutations", [&]() -> LocalWriteResult {
- // Load and apply all existing mutations. This lets us compute the current base state for
- // all non-idempotent transforms before applying any additional user-provided writes.
- MaybeDocumentMap existingDocuments = _localDocuments->GetDocuments(keys);
- // For non-idempotent mutations (such as `FieldValue.increment()`), we record the base
- // state in a separate patch mutation. This is later used to guarantee consistent values
- // and prevents flicker even if the backend sends us an update that already includes our
- // transform.
- std::vector<FSTMutation *> baseMutations;
- for (FSTMutation *mutation : mutations) {
- auto base_document_it = existingDocuments.find(mutation.key);
- FSTMaybeDocument *base_document =
- base_document_it != existingDocuments.end() ? base_document_it->second : nil;
- absl::optional<ObjectValue> base_value = [mutation extractBaseValue:base_document];
- if (base_value) {
- // NOTE: The base state should only be applied if there's some existing document to
- // override, so use a Precondition of exists=true
- baseMutations.push_back([[FSTPatchMutation alloc] initWithKey:mutation.key
- fieldMask:base_value->ToFieldMask()
- value:*base_value
- precondition:Precondition::Exists(true)]);
- }
- }
- FSTMutationBatch *batch = _mutationQueue->AddMutationBatch(
- localWriteTime, std::move(baseMutations), std::move(mutations));
- MaybeDocumentMap changedDocuments = [batch applyToLocalDocumentSet:existingDocuments];
- return LocalWriteResult{batch.batchID, std::move(changedDocuments)};
- });
- }
- - (MaybeDocumentMap)acknowledgeBatchWithResult:(FSTMutationBatchResult *)batchResult {
- return self.persistence.run("Acknowledge batch", [&]() -> MaybeDocumentMap {
- FSTMutationBatch *batch = batchResult.batch;
- _mutationQueue->AcknowledgeBatch(batch, batchResult.streamToken);
- [self applyBatchResult:batchResult];
- _mutationQueue->PerformConsistencyCheck();
- return _localDocuments->GetDocuments(batch.keys);
- });
- }
- - (MaybeDocumentMap)rejectBatchID:(BatchId)batchID {
- return self.persistence.run("Reject batch", [&]() -> MaybeDocumentMap {
- FSTMutationBatch *toReject = _mutationQueue->LookupMutationBatch(batchID);
- HARD_ASSERT(toReject, "Attempt to reject nonexistent batch!");
- _mutationQueue->RemoveMutationBatch(toReject);
- _mutationQueue->PerformConsistencyCheck();
- return _localDocuments->GetDocuments(toReject.keys);
- });
- }
- - (nullable NSData *)lastStreamToken {
- return _mutationQueue->GetLastStreamToken();
- }
- - (void)setLastStreamToken:(nullable NSData *)streamToken {
- self.persistence.run("Set stream token",
- [&]() { _mutationQueue->SetLastStreamToken(streamToken); });
- }
- - (const SnapshotVersion &)lastRemoteSnapshotVersion {
- return self.queryCache->GetLastRemoteSnapshotVersion();
- }
- - (MaybeDocumentMap)applyRemoteEvent:(const RemoteEvent &)remoteEvent {
- return self.persistence.run("Apply remote event", [&]() -> MaybeDocumentMap {
- // TODO(gsoltis): move the sequence number into the reference delegate.
- ListenSequenceNumber sequenceNumber = self.persistence.currentSequenceNumber;
- DocumentKeySet authoritativeUpdates;
- for (const auto &entry : remoteEvent.target_changes()) {
- TargetId targetID = entry.first;
- const TargetChange &change = entry.second;
- // Do not ref/unref unassigned targetIDs - it may lead to leaks.
- auto found = _targetIDs.find(targetID);
- if (found == _targetIDs.end()) {
- continue;
- }
- FSTQueryData *queryData = found->second;
- // When a global snapshot contains updates (either add or modify) we can completely trust
- // these updates as authoritative and blindly apply them to our cache (as a defensive measure
- // to promote self-healing in the unfortunate case that our cache is ever somehow corrupted /
- // out-of-sync).
- //
- // If the document is only updated while removing it from a target then watch isn't obligated
- // to send the absolute latest version: it can send the first version that caused the document
- // not to match.
- for (const DocumentKey &key : change.added_documents()) {
- authoritativeUpdates = authoritativeUpdates.insert(key);
- }
- for (const DocumentKey &key : change.modified_documents()) {
- authoritativeUpdates = authoritativeUpdates.insert(key);
- }
- _queryCache->RemoveMatchingKeys(change.removed_documents(), targetID);
- _queryCache->AddMatchingKeys(change.added_documents(), targetID);
- // Update the resume token if the change includes one. Don't clear any preexisting value.
- // Bump the sequence number as well, so that documents being removed now are ordered later
- // than documents that were previously removed from this target.
- NSData *resumeToken = change.resume_token();
- if (resumeToken.length > 0) {
- FSTQueryData *oldQueryData = queryData;
- queryData = [queryData queryDataByReplacingSnapshotVersion:remoteEvent.snapshot_version()
- resumeToken:resumeToken
- sequenceNumber:sequenceNumber];
- _targetIDs[targetID] = queryData;
- if ([self shouldPersistQueryData:queryData oldQueryData:oldQueryData change:change]) {
- _queryCache->UpdateTarget(queryData);
- }
- }
- }
- MaybeDocumentMap changedDocs;
- const DocumentKeySet &limboDocuments = remoteEvent.limbo_document_changes();
- DocumentKeySet updatedKeys;
- for (const auto &kv : remoteEvent.document_updates()) {
- updatedKeys = updatedKeys.insert(kv.first);
- }
- // Each loop iteration only affects its "own" doc, so it's safe to get all the remote
- // documents in advance in a single call.
- MaybeDocumentMap existingDocs = _remoteDocumentCache->GetAll(updatedKeys);
- for (const auto &kv : remoteEvent.document_updates()) {
- const DocumentKey &key = kv.first;
- FSTMaybeDocument *doc = kv.second;
- FSTMaybeDocument *existingDoc = nil;
- auto foundExisting = existingDocs.find(key);
- if (foundExisting != existingDocs.end()) {
- existingDoc = foundExisting->second;
- }
- // If a document update isn't authoritative, make sure we don't apply an old document version
- // to the remote cache. We make an exception for SnapshotVersion.MIN which can happen for
- // manufactured events (e.g. in the case of a limbo document resolution failing).
- if (!existingDoc || doc.version == SnapshotVersion::None() ||
- (authoritativeUpdates.contains(doc.key) && !existingDoc.hasPendingWrites) ||
- doc.version >= existingDoc.version) {
- _remoteDocumentCache->Add(doc);
- changedDocs = changedDocs.insert(key, doc);
- } else {
- LOG_DEBUG("FSTLocalStore Ignoring outdated watch update for %s. "
- "Current version: %s Watch version: %s",
- key.ToString(), existingDoc.version.timestamp().ToString(),
- doc.version.timestamp().ToString());
- }
- // If this was a limbo resolution, make sure we mark when it was accessed.
- if (limboDocuments.contains(key)) {
- [self.persistence.referenceDelegate limboDocumentUpdated:key];
- }
- }
- // HACK: The only reason we allow omitting snapshot version is so we can synthesize remote
- // events when we get permission denied errors while trying to resolve the state of a locally
- // cached document that is in limbo.
- const SnapshotVersion &lastRemoteVersion = _queryCache->GetLastRemoteSnapshotVersion();
- const SnapshotVersion &remoteVersion = remoteEvent.snapshot_version();
- if (remoteVersion != SnapshotVersion::None()) {
- HARD_ASSERT(remoteVersion >= lastRemoteVersion,
- "Watch stream reverted to previous snapshot?? (%s < %s)",
- remoteVersion.timestamp().ToString(), lastRemoteVersion.timestamp().ToString());
- _queryCache->SetLastRemoteSnapshotVersion(remoteVersion);
- }
- return _localDocuments->GetLocalViewOfDocuments(changedDocs);
- });
- }
- /**
- * Returns YES if the newQueryData should be persisted during an update of an active target.
- * QueryData should always be persisted when a target is being released and should not call this
- * function.
- *
- * While the target is active, QueryData updates can be omitted when nothing about the target has
- * changed except metadata like the resume token or snapshot version. Occasionally it's worth the
- * extra write to prevent these values from getting too stale after a crash, but this doesn't have
- * to be too frequent.
- */
- - (BOOL)shouldPersistQueryData:(FSTQueryData *)newQueryData
- oldQueryData:(FSTQueryData *)oldQueryData
- change:(const TargetChange &)change {
- // Avoid clearing any existing value
- if (newQueryData.resumeToken.length == 0) return NO;
- // Any resume token is interesting if there isn't one already.
- if (oldQueryData.resumeToken.length == 0) return YES;
- // Don't allow resume token changes to be buffered indefinitely. This allows us to be reasonably
- // up-to-date after a crash and avoids needing to loop over all active queries on shutdown.
- // Especially in the browser we may not get time to do anything interesting while the current
- // tab is closing.
- int64_t newSeconds = newQueryData.snapshotVersion.timestamp().seconds();
- int64_t oldSeconds = oldQueryData.snapshotVersion.timestamp().seconds();
- int64_t timeDelta = newSeconds - oldSeconds;
- if (timeDelta >= kResumeTokenMaxAgeSeconds) return YES;
- // Otherwise if the only thing that has changed about a target is its resume token then it's not
- // worth persisting. Note that the RemoteStore keeps an in-memory view of the currently active
- // targets which includes the current resume token, so stream failure or user changes will still
- // use an up-to-date resume token regardless of what we do here.
- size_t changes = change.added_documents().size() + change.modified_documents().size() +
- change.removed_documents().size();
- return changes > 0;
- }
- - (void)notifyLocalViewChanges:(const std::vector<LocalViewChanges> &)viewChanges {
- self.persistence.run("NotifyLocalViewChanges", [&]() {
- for (const LocalViewChanges &viewChange : viewChanges) {
- for (const DocumentKey &key : viewChange.removed_keys()) {
- [self->_persistence.referenceDelegate removeReference:key];
- }
- _localViewReferences.AddReferences(viewChange.added_keys(), viewChange.target_id());
- _localViewReferences.AddReferences(viewChange.removed_keys(), viewChange.target_id());
- }
- });
- }
- - (nullable FSTMutationBatch *)nextMutationBatchAfterBatchID:(BatchId)batchID {
- FSTMutationBatch *result =
- self.persistence.run("NextMutationBatchAfterBatchID", [&]() -> FSTMutationBatch * {
- return _mutationQueue->NextMutationBatchAfterBatchId(batchID);
- });
- return result;
- }
- - (nullable FSTMaybeDocument *)readDocument:(const DocumentKey &)key {
- return self.persistence.run("ReadDocument", [&]() -> FSTMaybeDocument *_Nullable {
- return _localDocuments->GetDocument(key);
- });
- }
- - (FSTQueryData *)allocateQuery:(FSTQuery *)query {
- FSTQueryData *queryData = self.persistence.run("Allocate query", [&]() -> FSTQueryData * {
- FSTQueryData *cached = _queryCache->GetTarget(query);
- // TODO(mcg): freshen last accessed date if cached exists?
- if (!cached) {
- cached = [[FSTQueryData alloc] initWithQuery:query
- targetID:_targetIDGenerator.NextId()
- listenSequenceNumber:self.persistence.currentSequenceNumber
- purpose:FSTQueryPurposeListen];
- _queryCache->AddTarget(cached);
- }
- return cached;
- });
- // Sanity check to ensure that even when resuming a query it's not currently active.
- TargetId targetID = queryData.targetID;
- HARD_ASSERT(_targetIDs.find(targetID) == _targetIDs.end(),
- "Tried to allocate an already allocated query: %s", query);
- _targetIDs[targetID] = queryData;
- return queryData;
- }
- - (void)releaseQuery:(FSTQuery *)query {
- self.persistence.run("Release query", [&]() {
- FSTQueryData *queryData = _queryCache->GetTarget(query);
- HARD_ASSERT(queryData, "Tried to release nonexistent query: %s", query);
- TargetId targetID = queryData.targetID;
- auto found = _targetIDs.find(targetID);
- if (found != _targetIDs.end()) {
- FSTQueryData *cachedQueryData = found->second;
- if (cachedQueryData.snapshotVersion > queryData.snapshotVersion) {
- // If we've been avoiding persisting the resumeToken (see shouldPersistQueryData for
- // conditions and rationale) we need to persist the token now because there will no
- // longer be an in-memory version to fall back on.
- queryData = cachedQueryData;
- _queryCache->UpdateTarget(queryData);
- }
- }
- // References for documents sent via Watch are automatically removed when we delete a
- // query's target data from the reference delegate. Since this does not remove references
- // for locally mutated documents, we have to remove the target associations for these
- // documents manually.
- DocumentKeySet removed = _localViewReferences.RemoveReferences(targetID);
- for (const DocumentKey &key : removed) {
- [self.persistence.referenceDelegate removeReference:key];
- }
- _targetIDs.erase(targetID);
- [self.persistence.referenceDelegate removeTarget:queryData];
- });
- }
- - (DocumentMap)executeQuery:(FSTQuery *)query {
- return self.persistence.run("ExecuteQuery", [&]() -> DocumentMap {
- return _localDocuments->GetDocumentsMatchingQuery(query);
- });
- }
- - (DocumentKeySet)remoteDocumentKeysForTarget:(TargetId)targetID {
- return self.persistence.run("RemoteDocumentKeysForTarget", [&]() -> DocumentKeySet {
- return _queryCache->GetMatchingKeys(targetID);
- });
- }
- - (void)applyBatchResult:(FSTMutationBatchResult *)batchResult {
- FSTMutationBatch *batch = batchResult.batch;
- DocumentKeySet docKeys = batch.keys;
- const DocumentVersionMap &versions = batchResult.docVersions;
- for (const DocumentKey &docKey : docKeys) {
- FSTMaybeDocument *_Nullable remoteDoc = _remoteDocumentCache->Get(docKey);
- FSTMaybeDocument *_Nullable doc = remoteDoc;
- auto ackVersionIter = versions.find(docKey);
- HARD_ASSERT(ackVersionIter != versions.end(),
- "docVersions should contain every doc in the write.");
- const SnapshotVersion &ackVersion = ackVersionIter->second;
- if (!doc || doc.version < ackVersion) {
- doc = [batch applyToRemoteDocument:doc documentKey:docKey mutationBatchResult:batchResult];
- if (!doc) {
- HARD_ASSERT(!remoteDoc, "Mutation batch %s applied to document %s resulted in nil.", batch,
- remoteDoc);
- } else {
- _remoteDocumentCache->Add(doc);
- }
- }
- }
- _mutationQueue->RemoveMutationBatch(batch);
- }
- - (LruResults)collectGarbage:(FSTLRUGarbageCollector *)garbageCollector {
- return self.persistence.run("Collect garbage", [&]() -> LruResults {
- return [garbageCollector collectWithLiveTargets:_targetIDs];
- });
- }
- @end
- NS_ASSUME_NONNULL_END
|