OLD | NEW |
| (Empty) |
1 // Copyright (c) 2012 The Chromium Authors. All rights reserved. | |
2 // Use of this source code is governed by a BSD-style license that can be | |
3 // found in the LICENSE file. | |
4 | |
5 #include "chrome/browser/sync/engine/syncer_util.h" | |
6 | |
7 #include <algorithm> | |
8 #include <set> | |
9 #include <string> | |
10 #include <vector> | |
11 | |
12 #include "base/location.h" | |
13 #include "base/metrics/histogram.h" | |
14 #include "chrome/browser/sync/engine/conflict_resolver.h" | |
15 #include "chrome/browser/sync/engine/nigori_util.h" | |
16 #include "chrome/browser/sync/engine/syncer_proto_util.h" | |
17 #include "chrome/browser/sync/engine/syncer_types.h" | |
18 #include "chrome/browser/sync/engine/syncproto.h" | |
19 #include "chrome/browser/sync/syncable/model_type.h" | |
20 #include "chrome/browser/sync/syncable/syncable.h" | |
21 #include "chrome/browser/sync/syncable/syncable_changes_version.h" | |
22 #include "chrome/browser/sync/util/cryptographer.h" | |
23 #include "chrome/browser/sync/util/time.h" | |
24 #include "sync/protocol/bookmark_specifics.pb.h" | |
25 #include "sync/protocol/nigori_specifics.pb.h" | |
26 #include "sync/protocol/password_specifics.pb.h" | |
27 #include "sync/protocol/sync.pb.h" | |
28 | |
29 using syncable::BASE_VERSION; | |
30 using syncable::Blob; | |
31 using syncable::CHANGES_VERSION; | |
32 using syncable::CREATE; | |
33 using syncable::CREATE_NEW_UPDATE_ITEM; | |
34 using syncable::CTIME; | |
35 using syncable::Directory; | |
36 using syncable::Entry; | |
37 using syncable::GetModelTypeFromSpecifics; | |
38 using syncable::GET_BY_HANDLE; | |
39 using syncable::GET_BY_ID; | |
40 using syncable::ID; | |
41 using syncable::IS_DEL; | |
42 using syncable::IS_DIR; | |
43 using syncable::IS_UNAPPLIED_UPDATE; | |
44 using syncable::IS_UNSYNCED; | |
45 using syncable::Id; | |
46 using syncable::IsRealDataType; | |
47 using syncable::META_HANDLE; | |
48 using syncable::MTIME; | |
49 using syncable::MutableEntry; | |
50 using syncable::NEXT_ID; | |
51 using syncable::NON_UNIQUE_NAME; | |
52 using syncable::BASE_SERVER_SPECIFICS; | |
53 using syncable::PARENT_ID; | |
54 using syncable::PREV_ID; | |
55 using syncable::ReadTransaction; | |
56 using syncable::SERVER_CTIME; | |
57 using syncable::SERVER_IS_DEL; | |
58 using syncable::SERVER_IS_DIR; | |
59 using syncable::SERVER_MTIME; | |
60 using syncable::SERVER_NON_UNIQUE_NAME; | |
61 using syncable::SERVER_PARENT_ID; | |
62 using syncable::SERVER_POSITION_IN_PARENT; | |
63 using syncable::SERVER_SPECIFICS; | |
64 using syncable::SERVER_VERSION; | |
65 using syncable::UNIQUE_CLIENT_TAG; | |
66 using syncable::UNIQUE_SERVER_TAG; | |
67 using syncable::SPECIFICS; | |
68 using syncable::SYNCER; | |
69 using syncable::WriteTransaction; | |
70 | |
71 namespace browser_sync { | |
72 | |
73 // Returns the number of unsynced entries. | |
74 // static | |
75 int SyncerUtil::GetUnsyncedEntries(syncable::BaseTransaction* trans, | |
76 std::vector<int64> *handles) { | |
77 trans->directory()->GetUnsyncedMetaHandles(trans, handles); | |
78 DVLOG_IF(1, !handles->empty()) << "Have " << handles->size() | |
79 << " unsynced items."; | |
80 return handles->size(); | |
81 } | |
82 | |
83 // static | |
84 void SyncerUtil::ChangeEntryIDAndUpdateChildren( | |
85 syncable::WriteTransaction* trans, | |
86 syncable::MutableEntry* entry, | |
87 const syncable::Id& new_id, | |
88 syncable::Directory::ChildHandles* children) { | |
89 syncable::Id old_id = entry->Get(ID); | |
90 if (!entry->Put(ID, new_id)) { | |
91 Entry old_entry(trans, GET_BY_ID, new_id); | |
92 CHECK(old_entry.good()); | |
93 LOG(FATAL) << "Attempt to change ID to " << new_id | |
94 << " conflicts with existing entry.\n\n" | |
95 << *entry << "\n\n" << old_entry; | |
96 } | |
97 if (entry->Get(IS_DIR)) { | |
98 // Get all child entries of the old id. | |
99 trans->directory()->GetChildHandlesById(trans, old_id, children); | |
100 Directory::ChildHandles::iterator i = children->begin(); | |
101 while (i != children->end()) { | |
102 MutableEntry child_entry(trans, GET_BY_HANDLE, *i++); | |
103 CHECK(child_entry.good()); | |
104 // Use the unchecked setter here to avoid touching the child's NEXT_ID | |
105 // and PREV_ID fields (which Put(PARENT_ID) would normally do to | |
106 // maintain linked-list invariants). In this case, NEXT_ID and PREV_ID | |
107 // among the children will be valid after the loop, since we update all | |
108 // the children at once. | |
109 child_entry.PutParentIdPropertyOnly(new_id); | |
110 } | |
111 } | |
112 // Update Id references on the previous and next nodes in the sibling | |
113 // order. Do this by reinserting into the linked list; the first | |
114 // step in PutPredecessor is to Unlink from the existing order, which | |
115 // will overwrite the stale Id value from the adjacent nodes. | |
116 if (entry->Get(PREV_ID) == entry->Get(NEXT_ID) && | |
117 entry->Get(PREV_ID) == old_id) { | |
118 // We just need a shallow update to |entry|'s fields since it is already | |
119 // self looped. | |
120 entry->Put(NEXT_ID, new_id); | |
121 entry->Put(PREV_ID, new_id); | |
122 } else { | |
123 entry->PutPredecessor(entry->Get(PREV_ID)); | |
124 } | |
125 } | |
126 | |
127 // static | |
128 void SyncerUtil::ChangeEntryIDAndUpdateChildren( | |
129 syncable::WriteTransaction* trans, | |
130 syncable::MutableEntry* entry, | |
131 const syncable::Id& new_id) { | |
132 syncable::Directory::ChildHandles children; | |
133 ChangeEntryIDAndUpdateChildren(trans, entry, new_id, &children); | |
134 } | |
135 | |
136 // static | |
137 syncable::Id SyncerUtil::FindLocalIdToUpdate( | |
138 syncable::BaseTransaction* trans, | |
139 const SyncEntity& update) { | |
140 // Expected entry points of this function: | |
141 // SyncEntity has NOT been applied to SERVER fields. | |
142 // SyncEntity has NOT been applied to LOCAL fields. | |
143 // DB has not yet been modified, no entries created for this update. | |
144 | |
145 const std::string& client_id = trans->directory()->cache_guid(); | |
146 | |
147 if (update.has_client_defined_unique_tag() && | |
148 !update.client_defined_unique_tag().empty()) { | |
149 // When a server sends down a client tag, the following cases can occur: | |
150 // 1) Client has entry for tag already, ID is server style, matches | |
151 // 2) Client has entry for tag already, ID is server, doesn't match. | |
152 // 3) Client has entry for tag already, ID is local, (never matches) | |
153 // 4) Client has no entry for tag | |
154 | |
155 // Case 1, we don't have to do anything since the update will | |
156 // work just fine. Update will end up in the proper entry, via ID lookup. | |
157 // Case 2 - Happens very rarely due to lax enforcement of client tags | |
158 // on the server, if two clients commit the same tag at the same time. | |
159 // When this happens, we pick the lexically-least ID and ignore all other | |
160 // items. | |
161 // Case 3 - We need to replace the local ID with the server ID so that | |
162 // this update gets targeted at the correct local entry; we expect conflict | |
163 // resolution to occur. | |
164 // Case 4 - Perfect. Same as case 1. | |
165 | |
166 syncable::Entry local_entry(trans, syncable::GET_BY_CLIENT_TAG, | |
167 update.client_defined_unique_tag()); | |
168 | |
169 // The SyncAPI equivalent of this function will return !good if IS_DEL. | |
170 // The syncable version will return good even if IS_DEL. | |
171 // TODO(chron): Unit test the case with IS_DEL and make sure. | |
172 if (local_entry.good()) { | |
173 if (local_entry.Get(ID).ServerKnows()) { | |
174 if (local_entry.Get(ID) != update.id()) { | |
175 // Case 2. | |
176 LOG(WARNING) << "Duplicated client tag."; | |
177 if (local_entry.Get(ID) < update.id()) { | |
178 // Signal an error; drop this update on the floor. Note that | |
179 // we don't server delete the item, because we don't allow it to | |
180 // exist locally at all. So the item will remain orphaned on | |
181 // the server, and we won't pay attention to it. | |
182 return syncable::GetNullId(); | |
183 } | |
184 } | |
185 // Target this change to the existing local entry; later, | |
186 // we'll change the ID of the local entry to update.id() | |
187 // if needed. | |
188 return local_entry.Get(ID); | |
189 } else { | |
190 // Case 3: We have a local entry with the same client tag. | |
191 // We should change the ID of the local entry to the server entry. | |
192 // This will result in an server ID with base version == 0, but that's | |
193 // a legal state for an item with a client tag. By changing the ID, | |
194 // update will now be applied to local_entry. | |
195 DCHECK(0 == local_entry.Get(BASE_VERSION) || | |
196 CHANGES_VERSION == local_entry.Get(BASE_VERSION)); | |
197 return local_entry.Get(ID); | |
198 } | |
199 } | |
200 } else if (update.has_originator_cache_guid() && | |
201 update.originator_cache_guid() == client_id) { | |
202 // If a commit succeeds, but the response does not come back fast enough | |
203 // then the syncer might assume that it was never committed. | |
204 // The server will track the client that sent up the original commit and | |
205 // return this in a get updates response. When this matches a local | |
206 // uncommitted item, we must mutate our local item and version to pick up | |
207 // the committed version of the same item whose commit response was lost. | |
208 // There is however still a race condition if the server has not | |
209 // completed the commit by the time the syncer tries to get updates | |
210 // again. To mitigate this, we need to have the server time out in | |
211 // a reasonable span, our commit batches have to be small enough | |
212 // to process within our HTTP response "assumed alive" time. | |
213 | |
214 // We need to check if we have an entry that didn't get its server | |
215 // id updated correctly. The server sends down a client ID | |
216 // and a local (negative) id. If we have a entry by that | |
217 // description, we should update the ID and version to the | |
218 // server side ones to avoid multiple copies of the same thing. | |
219 | |
220 syncable::Id client_item_id = syncable::Id::CreateFromClientString( | |
221 update.originator_client_item_id()); | |
222 DCHECK(!client_item_id.ServerKnows()); | |
223 syncable::Entry local_entry(trans, GET_BY_ID, client_item_id); | |
224 | |
225 // If it exists, then our local client lost a commit response. Use | |
226 // the local entry. | |
227 if (local_entry.good() && !local_entry.Get(IS_DEL)) { | |
228 int64 old_version = local_entry.Get(BASE_VERSION); | |
229 int64 new_version = update.version(); | |
230 DCHECK_LE(old_version, 0); | |
231 DCHECK_GT(new_version, 0); | |
232 // Otherwise setting the base version could cause a consistency failure. | |
233 // An entry should never be version 0 and SYNCED. | |
234 DCHECK(local_entry.Get(IS_UNSYNCED)); | |
235 | |
236 // Just a quick sanity check. | |
237 DCHECK(!local_entry.Get(ID).ServerKnows()); | |
238 | |
239 DVLOG(1) << "Reuniting lost commit response IDs. server id: " | |
240 << update.id() << " local id: " << local_entry.Get(ID) | |
241 << " new version: " << new_version; | |
242 | |
243 return local_entry.Get(ID); | |
244 } | |
245 } | |
246 // Fallback: target an entry having the server ID, creating one if needed. | |
247 return update.id(); | |
248 } | |
249 | |
250 // static | |
251 UpdateAttemptResponse SyncerUtil::AttemptToUpdateEntry( | |
252 syncable::WriteTransaction* const trans, | |
253 syncable::MutableEntry* const entry, | |
254 ConflictResolver* resolver, | |
255 Cryptographer* cryptographer) { | |
256 CHECK(entry->good()); | |
257 if (!entry->Get(IS_UNAPPLIED_UPDATE)) | |
258 return SUCCESS; // No work to do. | |
259 syncable::Id id = entry->Get(ID); | |
260 const sync_pb::EntitySpecifics& specifics = entry->Get(SERVER_SPECIFICS); | |
261 | |
262 // We intercept updates to the Nigori node, update the Cryptographer and | |
263 // encrypt any unsynced changes here because there is no Nigori | |
264 // ChangeProcessor. We never put the nigori node in a state of | |
265 // conflict_encryption. | |
266 // | |
267 // We always update the cryptographer with the server's nigori node, | |
268 // even if we have a locally modified nigori node (we manually merge nigori | |
269 // data in the conflict resolver in that case). This handles the case where | |
270 // two clients both set a different passphrase. The second client to attempt | |
271 // to commit will go into a state of having pending keys, unioned the set of | |
272 // encrypted types, and eventually re-encrypt everything with the passphrase | |
273 // of the first client and commit the set of merged encryption keys. Until the | |
274 // second client provides the pending passphrase, the cryptographer will | |
275 // preserve the encryption keys based on the local passphrase, while the | |
276 // nigori node will preserve the server encryption keys. | |
277 // | |
278 // If non-encryption changes are made to the nigori node, they will be | |
279 // lost as part of conflict resolution. This is intended, as we place a higher | |
280 // priority on preserving the server's passphrase change to preserving local | |
281 // non-encryption changes. Next time the non-encryption changes are made to | |
282 // the nigori node (e.g. on restart), they will commit without issue. | |
283 if (specifics.has_nigori()) { | |
284 const sync_pb::NigoriSpecifics& nigori = specifics.nigori(); | |
285 cryptographer->Update(nigori); | |
286 | |
287 // Make sure any unsynced changes are properly encrypted as necessary. | |
288 // We only perform this if the cryptographer is ready. If not, these are | |
289 // re-encrypted at SetPassphrase time (via ReEncryptEverything). This logic | |
290 // covers the case where the nigori updated marked new datatypes for | |
291 // encryption, but didn't change the passphrase. | |
292 if (cryptographer->is_ready()) { | |
293 // Note that we don't bother to encrypt any data for which IS_UNSYNCED | |
294 // == false here. The machine that turned on encryption should know about | |
295 // and re-encrypt all synced data. It's possible it could get interrupted | |
296 // during this process, but we currently reencrypt everything at startup | |
297 // as well, so as soon as a client is restarted with this datatype marked | |
298 // for encryption, all the data should be updated as necessary. | |
299 | |
300 // If this fails, something is wrong with the cryptographer, but there's | |
301 // nothing we can do about it here. | |
302 DVLOG(1) << "Received new nigori, encrypting unsynced changes."; | |
303 syncable::ProcessUnsyncedChangesForEncryption(trans, cryptographer); | |
304 } | |
305 } | |
306 | |
307 // Only apply updates that we can decrypt. If we can't decrypt the update, it | |
308 // is likely because the passphrase has not arrived yet. Because the | |
309 // passphrase may not arrive within this GetUpdates, we can't just return | |
310 // conflict, else we try to perform normal conflict resolution prematurely or | |
311 // the syncer may get stuck. As such, we return CONFLICT_ENCRYPTION, which is | |
312 // treated as an unresolvable conflict. See the description in syncer_types.h. | |
313 // This prevents any unsynced changes from commiting and postpones conflict | |
314 // resolution until all data can be decrypted. | |
315 if (specifics.has_encrypted() && | |
316 !cryptographer->CanDecrypt(specifics.encrypted())) { | |
317 // We can't decrypt this node yet. | |
318 DVLOG(1) << "Received an undecryptable " | |
319 << syncable::ModelTypeToString(entry->GetServerModelType()) | |
320 << " update, returning encryption_conflict."; | |
321 return CONFLICT_ENCRYPTION; | |
322 } else if (specifics.has_password() && | |
323 entry->Get(UNIQUE_SERVER_TAG).empty()) { | |
324 // Passwords use their own legacy encryption scheme. | |
325 const sync_pb::PasswordSpecifics& password = specifics.password(); | |
326 if (!cryptographer->CanDecrypt(password.encrypted())) { | |
327 DVLOG(1) << "Received an undecryptable password update, returning " | |
328 << "encryption_conflict."; | |
329 return CONFLICT_ENCRYPTION; | |
330 } | |
331 } | |
332 | |
333 if (!entry->Get(SERVER_IS_DEL)) { | |
334 syncable::Id new_parent = entry->Get(SERVER_PARENT_ID); | |
335 Entry parent(trans, GET_BY_ID, new_parent); | |
336 // A note on non-directory parents: | |
337 // We catch most unfixable tree invariant errors at update receipt time, | |
338 // however we deal with this case here because we may receive the child | |
339 // first then the illegal parent. Instead of dealing with it twice in | |
340 // different ways we deal with it once here to reduce the amount of code and | |
341 // potential errors. | |
342 if (!parent.good() || parent.Get(IS_DEL) || !parent.Get(IS_DIR)) { | |
343 return CONFLICT_HIERARCHY; | |
344 } | |
345 if (entry->Get(PARENT_ID) != new_parent) { | |
346 if (!entry->Get(IS_DEL) && !IsLegalNewParent(trans, id, new_parent)) { | |
347 DVLOG(1) << "Not updating item " << id | |
348 << ", illegal new parent (would cause loop)."; | |
349 return CONFLICT_HIERARCHY; | |
350 } | |
351 } | |
352 } else if (entry->Get(IS_DIR)) { | |
353 Directory::ChildHandles handles; | |
354 trans->directory()->GetChildHandlesById(trans, id, &handles); | |
355 if (!handles.empty()) { | |
356 // If we have still-existing children, then we need to deal with | |
357 // them before we can process this change. | |
358 DVLOG(1) << "Not deleting directory; it's not empty " << *entry; | |
359 return CONFLICT_HIERARCHY; | |
360 } | |
361 } | |
362 | |
363 if (entry->Get(IS_UNSYNCED)) { | |
364 DVLOG(1) << "Skipping update, returning conflict for: " << id | |
365 << " ; it's unsynced."; | |
366 return CONFLICT_SIMPLE; | |
367 } | |
368 | |
369 if (specifics.has_encrypted()) { | |
370 DVLOG(2) << "Received a decryptable " | |
371 << syncable::ModelTypeToString(entry->GetServerModelType()) | |
372 << " update, applying normally."; | |
373 } else { | |
374 DVLOG(2) << "Received an unencrypted " | |
375 << syncable::ModelTypeToString(entry->GetServerModelType()) | |
376 << " update, applying normally."; | |
377 } | |
378 | |
379 SyncerUtil::UpdateLocalDataFromServerData(trans, entry); | |
380 | |
381 return SUCCESS; | |
382 } | |
383 | |
384 namespace { | |
385 // Helper to synthesize a new-style sync_pb::EntitySpecifics for use locally, | |
386 // when the server speaks only the old sync_pb::SyncEntity_BookmarkData-based | |
387 // protocol. | |
388 void UpdateBookmarkSpecifics(const std::string& singleton_tag, | |
389 const std::string& url, | |
390 const std::string& favicon_bytes, | |
391 MutableEntry* local_entry) { | |
392 // In the new-style protocol, the server no longer sends bookmark info for | |
393 // the "google_chrome" folder. Mimic that here. | |
394 if (singleton_tag == "google_chrome") | |
395 return; | |
396 sync_pb::EntitySpecifics pb; | |
397 sync_pb::BookmarkSpecifics* bookmark = pb.mutable_bookmark(); | |
398 if (!url.empty()) | |
399 bookmark->set_url(url); | |
400 if (!favicon_bytes.empty()) | |
401 bookmark->set_favicon(favicon_bytes); | |
402 local_entry->Put(SERVER_SPECIFICS, pb); | |
403 } | |
404 | |
405 } // namespace | |
406 | |
407 // Pass in name and checksum because of UTF8 conversion. | |
408 // static | |
409 void SyncerUtil::UpdateServerFieldsFromUpdate( | |
410 MutableEntry* target, | |
411 const SyncEntity& update, | |
412 const std::string& name) { | |
413 if (update.deleted()) { | |
414 if (target->Get(SERVER_IS_DEL)) { | |
415 // If we already think the item is server-deleted, we're done. | |
416 // Skipping these cases prevents our committed deletions from coming | |
417 // back and overriding subsequent undeletions. For non-deleted items, | |
418 // the version number check has a similar effect. | |
419 return; | |
420 } | |
421 // The server returns very lightweight replies for deletions, so we don't | |
422 // clobber a bunch of fields on delete. | |
423 target->Put(SERVER_IS_DEL, true); | |
424 if (!target->Get(UNIQUE_CLIENT_TAG).empty()) { | |
425 // Items identified by the client unique tag are undeletable; when | |
426 // they're deleted, they go back to version 0. | |
427 target->Put(SERVER_VERSION, 0); | |
428 } else { | |
429 // Otherwise, fake a server version by bumping the local number. | |
430 target->Put(SERVER_VERSION, | |
431 std::max(target->Get(SERVER_VERSION), | |
432 target->Get(BASE_VERSION)) + 1); | |
433 } | |
434 target->Put(IS_UNAPPLIED_UPDATE, true); | |
435 return; | |
436 } | |
437 | |
438 DCHECK(target->Get(ID) == update.id()) | |
439 << "ID Changing not supported here"; | |
440 target->Put(SERVER_PARENT_ID, update.parent_id()); | |
441 target->Put(SERVER_NON_UNIQUE_NAME, name); | |
442 target->Put(SERVER_VERSION, update.version()); | |
443 target->Put(SERVER_CTIME, ProtoTimeToTime(update.ctime())); | |
444 target->Put(SERVER_MTIME, ProtoTimeToTime(update.mtime())); | |
445 target->Put(SERVER_IS_DIR, update.IsFolder()); | |
446 if (update.has_server_defined_unique_tag()) { | |
447 const std::string& tag = update.server_defined_unique_tag(); | |
448 target->Put(UNIQUE_SERVER_TAG, tag); | |
449 } | |
450 if (update.has_client_defined_unique_tag()) { | |
451 const std::string& tag = update.client_defined_unique_tag(); | |
452 target->Put(UNIQUE_CLIENT_TAG, tag); | |
453 } | |
454 // Store the datatype-specific part as a protobuf. | |
455 if (update.has_specifics()) { | |
456 DCHECK(update.GetModelType() != syncable::UNSPECIFIED) | |
457 << "Storing unrecognized datatype in sync database."; | |
458 target->Put(SERVER_SPECIFICS, update.specifics()); | |
459 } else if (update.has_bookmarkdata()) { | |
460 // Legacy protocol response for bookmark data. | |
461 const SyncEntity::BookmarkData& bookmark = update.bookmarkdata(); | |
462 UpdateBookmarkSpecifics(update.server_defined_unique_tag(), | |
463 bookmark.bookmark_url(), | |
464 bookmark.bookmark_favicon(), | |
465 target); | |
466 } | |
467 if (update.has_position_in_parent()) | |
468 target->Put(SERVER_POSITION_IN_PARENT, update.position_in_parent()); | |
469 | |
470 target->Put(SERVER_IS_DEL, update.deleted()); | |
471 // We only mark the entry as unapplied if its version is greater than the | |
472 // local data. If we're processing the update that corresponds to one of our | |
473 // commit we don't apply it as time differences may occur. | |
474 if (update.version() > target->Get(BASE_VERSION)) { | |
475 target->Put(IS_UNAPPLIED_UPDATE, true); | |
476 } | |
477 } | |
478 | |
479 // Creates a new Entry iff no Entry exists with the given id. | |
480 // static | |
481 void SyncerUtil::CreateNewEntry(syncable::WriteTransaction *trans, | |
482 const syncable::Id& id) { | |
483 syncable::MutableEntry entry(trans, GET_BY_ID, id); | |
484 if (!entry.good()) { | |
485 syncable::MutableEntry new_entry(trans, syncable::CREATE_NEW_UPDATE_ITEM, | |
486 id); | |
487 } | |
488 } | |
489 | |
490 // static | |
491 void SyncerUtil::SplitServerInformationIntoNewEntry( | |
492 syncable::WriteTransaction* trans, | |
493 syncable::MutableEntry* entry) { | |
494 syncable::Id id = entry->Get(ID); | |
495 ChangeEntryIDAndUpdateChildren(trans, entry, trans->directory()->NextId()); | |
496 entry->Put(BASE_VERSION, 0); | |
497 | |
498 MutableEntry new_entry(trans, CREATE_NEW_UPDATE_ITEM, id); | |
499 CopyServerFields(entry, &new_entry); | |
500 ClearServerData(entry); | |
501 | |
502 DVLOG(1) << "Splitting server information, local entry: " << *entry | |
503 << " server entry: " << new_entry; | |
504 } | |
505 | |
506 // This function is called on an entry when we can update the user-facing data | |
507 // from the server data. | |
508 // static | |
509 void SyncerUtil::UpdateLocalDataFromServerData( | |
510 syncable::WriteTransaction* trans, | |
511 syncable::MutableEntry* entry) { | |
512 DCHECK(!entry->Get(IS_UNSYNCED)); | |
513 DCHECK(entry->Get(IS_UNAPPLIED_UPDATE)); | |
514 | |
515 DVLOG(2) << "Updating entry : " << *entry; | |
516 // Start by setting the properties that determine the model_type. | |
517 entry->Put(SPECIFICS, entry->Get(SERVER_SPECIFICS)); | |
518 // Clear the previous server specifics now that we're applying successfully. | |
519 entry->Put(BASE_SERVER_SPECIFICS, sync_pb::EntitySpecifics()); | |
520 entry->Put(IS_DIR, entry->Get(SERVER_IS_DIR)); | |
521 // This strange dance around the IS_DEL flag avoids problems when setting | |
522 // the name. | |
523 // TODO(chron): Is this still an issue? Unit test this codepath. | |
524 if (entry->Get(SERVER_IS_DEL)) { | |
525 entry->Put(IS_DEL, true); | |
526 } else { | |
527 entry->Put(NON_UNIQUE_NAME, entry->Get(SERVER_NON_UNIQUE_NAME)); | |
528 entry->Put(PARENT_ID, entry->Get(SERVER_PARENT_ID)); | |
529 CHECK(entry->Put(IS_DEL, false)); | |
530 Id new_predecessor = | |
531 entry->ComputePrevIdFromServerPosition(entry->Get(SERVER_PARENT_ID)); | |
532 CHECK(entry->PutPredecessor(new_predecessor)) | |
533 << " Illegal predecessor after converting from server position."; | |
534 } | |
535 | |
536 entry->Put(CTIME, entry->Get(SERVER_CTIME)); | |
537 entry->Put(MTIME, entry->Get(SERVER_MTIME)); | |
538 entry->Put(BASE_VERSION, entry->Get(SERVER_VERSION)); | |
539 entry->Put(IS_DEL, entry->Get(SERVER_IS_DEL)); | |
540 entry->Put(IS_UNAPPLIED_UPDATE, false); | |
541 } | |
542 | |
543 // static | |
544 VerifyCommitResult SyncerUtil::ValidateCommitEntry( | |
545 syncable::Entry* entry) { | |
546 syncable::Id id = entry->Get(ID); | |
547 if (id == entry->Get(PARENT_ID)) { | |
548 CHECK(id.IsRoot()) << "Non-root item is self parenting." << *entry; | |
549 // If the root becomes unsynced it can cause us problems. | |
550 LOG(ERROR) << "Root item became unsynced " << *entry; | |
551 return VERIFY_UNSYNCABLE; | |
552 } | |
553 if (entry->IsRoot()) { | |
554 LOG(ERROR) << "Permanent item became unsynced " << *entry; | |
555 return VERIFY_UNSYNCABLE; | |
556 } | |
557 if (entry->Get(IS_DEL) && !entry->Get(ID).ServerKnows()) { | |
558 // Drop deleted uncommitted entries. | |
559 return VERIFY_UNSYNCABLE; | |
560 } | |
561 return VERIFY_OK; | |
562 } | |
563 | |
564 // static | |
565 bool SyncerUtil::AddItemThenPredecessors( | |
566 syncable::BaseTransaction* trans, | |
567 syncable::Entry* item, | |
568 syncable::IndexedBitField inclusion_filter, | |
569 syncable::MetahandleSet* inserted_items, | |
570 std::vector<syncable::Id>* commit_ids) { | |
571 | |
572 if (!inserted_items->insert(item->Get(META_HANDLE)).second) | |
573 return false; | |
574 commit_ids->push_back(item->Get(ID)); | |
575 if (item->Get(IS_DEL)) | |
576 return true; // Deleted items have no predecessors. | |
577 | |
578 Id prev_id = item->Get(PREV_ID); | |
579 while (!prev_id.IsRoot()) { | |
580 Entry prev(trans, GET_BY_ID, prev_id); | |
581 CHECK(prev.good()) << "Bad id when walking predecessors."; | |
582 if (!prev.Get(inclusion_filter)) | |
583 break; | |
584 if (!inserted_items->insert(prev.Get(META_HANDLE)).second) | |
585 break; | |
586 commit_ids->push_back(prev_id); | |
587 prev_id = prev.Get(PREV_ID); | |
588 } | |
589 return true; | |
590 } | |
591 | |
592 // static | |
593 void SyncerUtil::AddPredecessorsThenItem( | |
594 syncable::BaseTransaction* trans, | |
595 syncable::Entry* item, | |
596 syncable::IndexedBitField inclusion_filter, | |
597 syncable::MetahandleSet* inserted_items, | |
598 std::vector<syncable::Id>* commit_ids) { | |
599 size_t initial_size = commit_ids->size(); | |
600 if (!AddItemThenPredecessors(trans, item, inclusion_filter, inserted_items, | |
601 commit_ids)) | |
602 return; | |
603 // Reverse what we added to get the correct order. | |
604 std::reverse(commit_ids->begin() + initial_size, commit_ids->end()); | |
605 } | |
606 | |
607 // static | |
608 void SyncerUtil::MarkDeletedChildrenSynced( | |
609 syncable::Directory* dir, | |
610 std::set<syncable::Id>* deleted_folders) { | |
611 // There's two options here. | |
612 // 1. Scan deleted unsynced entries looking up their pre-delete tree for any | |
613 // of the deleted folders. | |
614 // 2. Take each folder and do a tree walk of all entries underneath it. | |
615 // #2 has a lower big O cost, but writing code to limit the time spent inside | |
616 // the transaction during each step is simpler with 1. Changing this decision | |
617 // may be sensible if this code shows up in profiling. | |
618 if (deleted_folders->empty()) | |
619 return; | |
620 Directory::UnsyncedMetaHandles handles; | |
621 { | |
622 ReadTransaction trans(FROM_HERE, dir); | |
623 dir->GetUnsyncedMetaHandles(&trans, &handles); | |
624 } | |
625 if (handles.empty()) | |
626 return; | |
627 Directory::UnsyncedMetaHandles::iterator it; | |
628 for (it = handles.begin() ; it != handles.end() ; ++it) { | |
629 // Single transaction / entry we deal with. | |
630 WriteTransaction trans(FROM_HERE, SYNCER, dir); | |
631 MutableEntry entry(&trans, GET_BY_HANDLE, *it); | |
632 if (!entry.Get(IS_UNSYNCED) || !entry.Get(IS_DEL)) | |
633 continue; | |
634 syncable::Id id = entry.Get(PARENT_ID); | |
635 while (id != trans.root_id()) { | |
636 if (deleted_folders->find(id) != deleted_folders->end()) { | |
637 // We've synced the deletion of this deleted entries parent. | |
638 entry.Put(IS_UNSYNCED, false); | |
639 break; | |
640 } | |
641 Entry parent(&trans, GET_BY_ID, id); | |
642 if (!parent.good() || !parent.Get(IS_DEL)) | |
643 break; | |
644 id = parent.Get(PARENT_ID); | |
645 } | |
646 } | |
647 } | |
648 | |
649 // static | |
650 VerifyResult SyncerUtil::VerifyNewEntry( | |
651 const SyncEntity& update, | |
652 syncable::Entry* target, | |
653 const bool deleted) { | |
654 if (target->good()) { | |
655 // Not a new update. | |
656 return VERIFY_UNDECIDED; | |
657 } | |
658 if (deleted) { | |
659 // Deletion of an item we've never seen can be ignored. | |
660 return VERIFY_SKIP; | |
661 } | |
662 | |
663 return VERIFY_SUCCESS; | |
664 } | |
665 | |
666 // Assumes we have an existing entry; check here for updates that break | |
667 // consistency rules. | |
668 // static | |
669 VerifyResult SyncerUtil::VerifyUpdateConsistency( | |
670 syncable::WriteTransaction* trans, | |
671 const SyncEntity& update, | |
672 syncable::MutableEntry* target, | |
673 const bool deleted, | |
674 const bool is_directory, | |
675 syncable::ModelType model_type) { | |
676 | |
677 CHECK(target->good()); | |
678 | |
679 // If the update is a delete, we don't really need to worry at this stage. | |
680 if (deleted) | |
681 return VERIFY_SUCCESS; | |
682 | |
683 if (model_type == syncable::UNSPECIFIED) { | |
684 // This update is to an item of a datatype we don't recognize. The server | |
685 // shouldn't have sent it to us. Throw it on the ground. | |
686 return VERIFY_SKIP; | |
687 } | |
688 | |
689 if (target->Get(SERVER_VERSION) > 0) { | |
690 // Then we've had an update for this entry before. | |
691 if (is_directory != target->Get(SERVER_IS_DIR) || | |
692 model_type != target->GetServerModelType()) { | |
693 if (target->Get(IS_DEL)) { // If we've deleted the item, we don't care. | |
694 return VERIFY_SKIP; | |
695 } else { | |
696 LOG(ERROR) << "Server update doesn't agree with previous updates. "; | |
697 LOG(ERROR) << " Entry: " << *target; | |
698 LOG(ERROR) << " Update: " | |
699 << SyncerProtoUtil::SyncEntityDebugString(update); | |
700 return VERIFY_FAIL; | |
701 } | |
702 } | |
703 | |
704 if (!deleted && (target->Get(ID) == update.id()) && | |
705 (target->Get(SERVER_IS_DEL) || | |
706 (!target->Get(IS_UNSYNCED) && target->Get(IS_DEL) && | |
707 target->Get(BASE_VERSION) > 0))) { | |
708 // An undelete. The latter case in the above condition is for | |
709 // when the server does not give us an update following the | |
710 // commit of a delete, before undeleting. | |
711 // Undeletion is common for items that reuse the client-unique tag. | |
712 VerifyResult result = | |
713 SyncerUtil::VerifyUndelete(trans, update, target); | |
714 if (VERIFY_UNDECIDED != result) | |
715 return result; | |
716 } | |
717 } | |
718 if (target->Get(BASE_VERSION) > 0) { | |
719 // We've committed this update in the past. | |
720 if (is_directory != target->Get(IS_DIR) || | |
721 model_type != target->GetModelType()) { | |
722 LOG(ERROR) << "Server update doesn't agree with committed item. "; | |
723 LOG(ERROR) << " Entry: " << *target; | |
724 LOG(ERROR) << " Update: " | |
725 << SyncerProtoUtil::SyncEntityDebugString(update); | |
726 return VERIFY_FAIL; | |
727 } | |
728 if (target->Get(ID) == update.id()) { | |
729 if (target->Get(SERVER_VERSION) > update.version()) { | |
730 LOG(WARNING) << "We've already seen a more recent version."; | |
731 LOG(WARNING) << " Entry: " << *target; | |
732 LOG(WARNING) << " Update: " | |
733 << SyncerProtoUtil::SyncEntityDebugString(update); | |
734 return VERIFY_SKIP; | |
735 } | |
736 } | |
737 } | |
738 return VERIFY_SUCCESS; | |
739 } | |
740 | |
741 // Assumes we have an existing entry; verify an update that seems to be | |
742 // expressing an 'undelete' | |
743 // static | |
744 VerifyResult SyncerUtil::VerifyUndelete(syncable::WriteTransaction* trans, | |
745 const SyncEntity& update, | |
746 syncable::MutableEntry* target) { | |
747 // TODO(nick): We hit this path for items deleted items that the server | |
748 // tells us to re-create; only deleted items with positive base versions | |
749 // will hit this path. However, it's not clear how such an undeletion | |
750 // would actually succeed on the server; in the protocol, a base | |
751 // version of 0 is required to undelete an object. This codepath | |
752 // should be deprecated in favor of client-tag style undeletion | |
753 // (where items go to version 0 when they're deleted), or else | |
754 // removed entirely (if this type of undeletion is indeed impossible). | |
755 CHECK(target->good()); | |
756 DVLOG(1) << "Server update is attempting undelete. " << *target | |
757 << "Update:" << SyncerProtoUtil::SyncEntityDebugString(update); | |
758 // Move the old one aside and start over. It's too tricky to get the old one | |
759 // back into a state that would pass CheckTreeInvariants(). | |
760 if (target->Get(IS_DEL)) { | |
761 DCHECK(target->Get(UNIQUE_CLIENT_TAG).empty()) | |
762 << "Doing move-aside undeletion on client-tagged item."; | |
763 target->Put(ID, trans->directory()->NextId()); | |
764 target->Put(UNIQUE_CLIENT_TAG, ""); | |
765 target->Put(BASE_VERSION, CHANGES_VERSION); | |
766 target->Put(SERVER_VERSION, 0); | |
767 return VERIFY_SUCCESS; | |
768 } | |
769 if (update.version() < target->Get(SERVER_VERSION)) { | |
770 LOG(WARNING) << "Update older than current server version for " | |
771 << *target << " Update:" | |
772 << SyncerProtoUtil::SyncEntityDebugString(update); | |
773 return VERIFY_SUCCESS; // Expected in new sync protocol. | |
774 } | |
775 return VERIFY_UNDECIDED; | |
776 } | |
777 | |
778 } // namespace browser_sync | |
OLD | NEW |