mirror of
https://gitlab.winehq.org/wine/wine-gecko.git
synced 2024-09-13 09:24:08 -07:00
553 lines
18 KiB
JavaScript
553 lines
18 KiB
JavaScript
/* ***** BEGIN LICENSE BLOCK *****
|
|
* Version: MPL 1.1/GPL 2.0/LGPL 2.1
|
|
*
|
|
* The contents of this file are subject to the Mozilla Public License Version
|
|
* 1.1 (the "License"); you may not use this file except in compliance with
|
|
* the License. You may obtain a copy of the License at
|
|
* http://www.mozilla.org/MPL/
|
|
*
|
|
* Software distributed under the License is distributed on an "AS IS" basis,
|
|
* WITHOUT WARRANTY OF ANY KIND, either express or implied. See the License
|
|
* for the specific language governing rights and limitations under the
|
|
* License.
|
|
*
|
|
* The Original Code is Bookmarks Sync.
|
|
*
|
|
* The Initial Developer of the Original Code is Mozilla.
|
|
* Portions created by the Initial Developer are Copyright (C) 2007
|
|
* the Initial Developer. All Rights Reserved.
|
|
*
|
|
* Contributor(s):
|
|
* Dan Mills <thunder@mozilla.com>
|
|
* Myk Melez <myk@mozilla.org>
|
|
*
|
|
* Alternatively, the contents of this file may be used under the terms of
|
|
* either the GNU General Public License Version 2 or later (the "GPL"), or
|
|
* the GNU Lesser General Public License Version 2.1 or later (the "LGPL"),
|
|
* in which case the provisions of the GPL or the LGPL are applicable instead
|
|
* of those above. If you wish to allow use of your version of this file only
|
|
* under the terms of either the GPL or the LGPL, and not to allow others to
|
|
* use your version of this file under the terms of the MPL, indicate your
|
|
* decision by deleting the provisions above and replace them with the notice
|
|
* and other provisions required by the GPL or the LGPL. If you do not delete
|
|
* the provisions above, a recipient may use your version of this file under
|
|
* the terms of any one of the MPL, the GPL or the LGPL.
|
|
*
|
|
* ***** END LICENSE BLOCK ***** */
|
|
|
|
const EXPORTED_SYMBOLS = ['Engines', 'Engine', 'SyncEngine'];
|
|
|
|
const Cc = Components.classes;
|
|
const Ci = Components.interfaces;
|
|
const Cr = Components.results;
|
|
const Cu = Components.utils;
|
|
|
|
Cu.import("resource://gre/modules/XPCOMUtils.jsm");
|
|
Cu.import("resource://weave/ext/Observers.js");
|
|
Cu.import("resource://weave/log4moz.js");
|
|
Cu.import("resource://weave/constants.js");
|
|
Cu.import("resource://weave/util.js");
|
|
Cu.import("resource://weave/wrap.js");
|
|
Cu.import("resource://weave/resource.js");
|
|
Cu.import("resource://weave/identity.js");
|
|
Cu.import("resource://weave/stores.js");
|
|
Cu.import("resource://weave/trackers.js");
|
|
Cu.import("resource://weave/async.js");
|
|
|
|
Cu.import("resource://weave/base_records/wbo.js");
|
|
Cu.import("resource://weave/base_records/keys.js");
|
|
Cu.import("resource://weave/base_records/crypto.js");
|
|
Cu.import("resource://weave/base_records/collection.js");
|
|
|
|
Function.prototype.async = Async.sugar;
|
|
|
|
// Singleton service, holds registered engines
|
|
|
|
Utils.lazy(this, 'Engines', EngineManagerSvc);
|
|
|
|
function EngineManagerSvc() {
|
|
this._engines = {};
|
|
this._log = Log4Moz.repository.getLogger("Service.Engines");
|
|
this._log.level = Log4Moz.Level[Svc.Prefs.get(
|
|
"log.logger.service.engines", "Debug")];
|
|
}
|
|
EngineManagerSvc.prototype = {
|
|
get: function EngMgr_get(name) {
|
|
// Return an array of engines if we have an array of names
|
|
if (Utils.isArray(name)) {
|
|
let engines = [];
|
|
name.forEach(function(name) {
|
|
let engine = this.get(name);
|
|
if (engine)
|
|
engines.push(engine);
|
|
}, this);
|
|
return engines;
|
|
}
|
|
|
|
let engine = this._engines[name];
|
|
if (!engine)
|
|
this._log.debug("Could not get engine: " + name);
|
|
return engine;
|
|
},
|
|
getAll: function EngMgr_getAll() {
|
|
let ret = [];
|
|
for (key in this._engines) {
|
|
ret.push(this._engines[key]);
|
|
}
|
|
return ret;
|
|
},
|
|
getEnabled: function EngMgr_getEnabled() {
|
|
let ret = [];
|
|
for (key in this._engines) {
|
|
if(this._engines[key].enabled)
|
|
ret.push(this._engines[key]);
|
|
}
|
|
return ret;
|
|
},
|
|
|
|
/**
|
|
* Register an Engine to the service. Alternatively, give an array of engine
|
|
* objects to register.
|
|
*
|
|
* @param engineObject
|
|
* Engine object used to get an instance of the engine
|
|
* @return The engine object if anything failed
|
|
*/
|
|
register: function EngMgr_register(engineObject) {
|
|
if (Utils.isArray(engineObject))
|
|
return engineObject.map(this.register, this);
|
|
|
|
try {
|
|
let name = engineObject.prototype.name;
|
|
if (name in this._engines)
|
|
this._log.error("Engine '" + name + "' is already registered!");
|
|
else
|
|
this._engines[name] = new engineObject();
|
|
}
|
|
catch(ex) {
|
|
let mesg = ex.message ? ex.message : ex;
|
|
let name = engineObject || "";
|
|
name = name.prototype || "";
|
|
name = name.name || "";
|
|
|
|
let out = "Could not initialize engine '" + name + "': " + mesg;
|
|
dump(out);
|
|
this._log.error(out);
|
|
|
|
return engineObject;
|
|
}
|
|
},
|
|
unregister: function EngMgr_unregister(val) {
|
|
let name = val;
|
|
if (val instanceof Engine)
|
|
name = val.name;
|
|
delete this._engines[name];
|
|
}
|
|
};
|
|
|
|
function Engine() { this._init(); }
|
|
Engine.prototype = {
|
|
name: "engine",
|
|
displayName: "Boring Engine",
|
|
logName: "Engine",
|
|
|
|
// _storeObj, and _trackerObj should to be overridden in subclasses
|
|
|
|
_storeObj: Store,
|
|
_trackerObj: Tracker,
|
|
|
|
get enabled() Utils.prefs.getBoolPref("engine." + this.name),
|
|
get score() this._tracker.score,
|
|
|
|
get _store() {
|
|
if (!this.__store)
|
|
this.__store = new this._storeObj();
|
|
return this.__store;
|
|
},
|
|
|
|
get _tracker() {
|
|
if (!this.__tracker)
|
|
this.__tracker = new this._trackerObj();
|
|
return this.__tracker;
|
|
},
|
|
|
|
_init: function Engine__init() {
|
|
let levelPref = "log.logger.engine." + this.name;
|
|
let level = "Debug";
|
|
try { level = Utils.prefs.getCharPref(levelPref); }
|
|
catch (e) { /* ignore unset prefs */ }
|
|
|
|
this._notify = Wrap.notify("weave:engine:");
|
|
this._log = Log4Moz.repository.getLogger("Engine." + this.logName);
|
|
this._log.level = Log4Moz.Level[level];
|
|
|
|
this._tracker; // initialize tracker to load previously changed IDs
|
|
this._log.debug("Engine initialized");
|
|
},
|
|
|
|
sync: function Engine_sync(onComplete) {
|
|
if (!this._sync)
|
|
throw "engine does not implement _sync method";
|
|
this._notify("sync", this.name, this._sync).async(this, onComplete);
|
|
},
|
|
|
|
wipeServer: function Engimne_wipeServer(onComplete) {
|
|
if (!this._wipeServer)
|
|
throw "engine does not implement _wipeServer method";
|
|
this._notify("wipe-server", this.name, this._wipeServer).async(this, onComplete);
|
|
},
|
|
|
|
/**
|
|
* Get rid of any local meta-data
|
|
*/
|
|
resetClient: function Engine_resetClient(onComplete) {
|
|
if (!this._resetClient)
|
|
throw "engine does not implement _resetClient method";
|
|
|
|
this._notify("reset-client", this.name, this._resetClient).
|
|
async(this, onComplete);
|
|
},
|
|
|
|
_wipeClient: function Engine__wipeClient() {
|
|
let self = yield;
|
|
|
|
yield this.resetClient(self.cb);
|
|
|
|
this._log.debug("Deleting all local data");
|
|
this._store.wipe();
|
|
},
|
|
|
|
wipeClient: function Engine_wipeClient(onComplete) {
|
|
this._notify("wipe-client", this.name, this._wipeClient).async(this, onComplete);
|
|
}
|
|
};
|
|
|
|
function SyncEngine() { this._init(); }
|
|
SyncEngine.prototype = {
|
|
__proto__: Engine.prototype,
|
|
|
|
_recordObj: CryptoWrapper,
|
|
|
|
get baseURL() {
|
|
let url = Svc.Prefs.get("clusterURL");
|
|
if (!url)
|
|
return null;
|
|
if (url[url.length-1] != '/')
|
|
url += '/';
|
|
url += "0.3/user/";
|
|
return url;
|
|
},
|
|
|
|
get engineURL() {
|
|
return this.baseURL + ID.get('WeaveID').username + '/' + this.name + '/';
|
|
},
|
|
|
|
get cryptoMetaURL() {
|
|
return this.baseURL + ID.get('WeaveID').username + '/crypto/' + this.name;
|
|
},
|
|
|
|
get lastSync() {
|
|
return Svc.Prefs.get(this.name + ".lastSync", 0);
|
|
},
|
|
set lastSync(value) {
|
|
Svc.Prefs.reset(this.name + ".lastSync");
|
|
if (typeof(value) == "string")
|
|
value = parseInt(value);
|
|
Svc.Prefs.set(this.name + ".lastSync", value);
|
|
},
|
|
resetLastSync: function SyncEngine_resetLastSync() {
|
|
this._log.debug("Resetting " + this.name + " last sync time");
|
|
Svc.Prefs.reset(this.name + ".lastSync");
|
|
Svc.Prefs.set(this.name + ".lastSync", 0);
|
|
},
|
|
|
|
// Create a new record by querying the store, and add the engine metadata
|
|
_createRecord: function SyncEngine__createRecord(id) {
|
|
return this._store.createRecord(id, this.cryptoMetaURL);
|
|
},
|
|
|
|
// Check if a record is "like" another one, even though the IDs are different,
|
|
// in that case, we'll change the ID of the local item to match
|
|
// Probably needs to be overridden in a subclass, to change which criteria
|
|
// make two records "the same one"
|
|
_recordLike: function SyncEngine__recordLike(a, b) {
|
|
if (a.parentid != b.parentid)
|
|
return false;
|
|
if (a.depth != b.depth)
|
|
return false;
|
|
// note: sortindex ignored
|
|
if (a.deleted || b.deleted)
|
|
return false;
|
|
return Utils.deepEquals(a.cleartext, b.cleartext);
|
|
},
|
|
|
|
_lowMemCheck: function SyncEngine__lowMemCheck() {
|
|
if (Svc.Memory.isLowMemory()) {
|
|
this._log.warn("Low memory, forcing GC");
|
|
Cu.forceGC();
|
|
if (Svc.Memory.isLowMemory()) {
|
|
this._log.warn("Low memory, aborting sync!");
|
|
throw "Low memory";
|
|
}
|
|
}
|
|
},
|
|
|
|
// Any setup that needs to happen at the beginning of each sync.
|
|
// Makes sure crypto records and keys are all set-up
|
|
_syncStartup: function SyncEngine__syncStartup() {
|
|
let self = yield;
|
|
|
|
this._log.debug("Ensuring server crypto records are there");
|
|
|
|
let meta = yield CryptoMetas.get(self.cb, this.cryptoMetaURL);
|
|
if (!meta) {
|
|
let symkey = Svc.Crypto.generateRandomKey();
|
|
let pubkey = yield PubKeys.getDefaultKey(self.cb);
|
|
meta = new CryptoMeta(this.cryptoMetaURL);
|
|
meta.generateIV();
|
|
yield meta.addUnwrappedKey(self.cb, pubkey, symkey);
|
|
let res = new Resource(meta.uri);
|
|
yield res.put(self.cb, meta.serialize());
|
|
}
|
|
|
|
// first sync special case: upload all items
|
|
// NOTE: we use a backdoor (of sorts) to the tracker so it
|
|
// won't save to disk this list over and over
|
|
if (!this.lastSync) {
|
|
this._log.info("First sync, uploading all items");
|
|
this._tracker.clearChangedIDs();
|
|
[i for (i in this._store.getAllIDs())]
|
|
.forEach(function(id) this._tracker.changedIDs[id] = true, this);
|
|
}
|
|
|
|
let outnum = [i for (i in this._tracker.changedIDs)].length;
|
|
this._log.info(outnum + " outgoing items pre-reconciliation");
|
|
},
|
|
|
|
// Generate outgoing records
|
|
_processIncoming: function SyncEngine__processIncoming() {
|
|
let self = yield;
|
|
|
|
this._log.debug("Downloading & applying server changes");
|
|
|
|
// enable cache, and keep only the first few items. Otherwise (when
|
|
// we have more outgoing items than can fit in the cache), we will
|
|
// keep rotating items in and out, perpetually getting cache misses
|
|
this._store.cache.enabled = true;
|
|
this._store.cache.fifo = false; // filo
|
|
this._store.cache.clear();
|
|
|
|
let newitems = new Collection(this.engineURL, this._recordObj);
|
|
newitems.newer = this.lastSync;
|
|
newitems.full = true;
|
|
newitems.sort = "depthindex";
|
|
yield newitems.get(self.cb);
|
|
|
|
let item;
|
|
let count = {applied: 0, reconciled: 0};
|
|
this._lastSyncTmp = 0;
|
|
|
|
while ((item = yield newitems.iter.next(self.cb))) {
|
|
this._lowMemCheck();
|
|
try {
|
|
yield item.decrypt(self.cb, ID.get('WeaveCryptoID').password);
|
|
if (yield this._reconcile.async(this, self.cb, item)) {
|
|
count.applied++;
|
|
yield this._applyIncoming.async(this, self.cb, item);
|
|
} else {
|
|
count.reconciled++;
|
|
this._log.trace("Skipping reconciled incoming item " + item.id);
|
|
if (this._lastSyncTmp < item.modified)
|
|
this._lastSyncTmp = item.modified;
|
|
}
|
|
} catch (e) {
|
|
this._log.error("Could not process incoming record: " +
|
|
Utils.exceptionStr(e));
|
|
}
|
|
}
|
|
if (this.lastSync < this._lastSyncTmp)
|
|
this.lastSync = this._lastSyncTmp;
|
|
|
|
this._log.info("Applied " + count.applied + " records, reconciled " +
|
|
count.reconciled + " records");
|
|
|
|
// try to free some memory
|
|
this._store.cache.clear();
|
|
Cu.forceGC();
|
|
},
|
|
|
|
_isEqual: function SyncEngine__isEqual(item) {
|
|
let local = this._createRecord(item.id);
|
|
this._log.trace("Local record: \n" + local);
|
|
if (item.parentid == local.parentid &&
|
|
item.sortindex == local.sortindex &&
|
|
Utils.deepEquals(item.cleartext, local.cleartext)) {
|
|
this._log.trace("Local record is the same");
|
|
return true;
|
|
} else {
|
|
this._log.trace("Local record is different");
|
|
return false;
|
|
}
|
|
},
|
|
|
|
// Reconciliation has three steps:
|
|
// 1) Check for the same item (same ID) on both the incoming and outgoing
|
|
// queues. This means the same item was modified on this profile and
|
|
// another at the same time. In this case, this client wins (which really
|
|
// means, the last profile you sync wins).
|
|
// 2) Check if the incoming item's ID exists locally. In that case it's an
|
|
// update and we should not try a similarity check (step 3)
|
|
// 3) Check if any incoming & outgoing items are actually the same, even
|
|
// though they have different IDs. This happens when the same item is
|
|
// added on two different machines at the same time. It's also the common
|
|
// case when syncing for the first time two machines that already have the
|
|
// same bookmarks. In this case we change the IDs to match.
|
|
_reconcile: function SyncEngine__reconcile(item) {
|
|
let self = yield;
|
|
let ret = true;
|
|
|
|
// Step 1: Check for conflicts
|
|
// If same as local record, do not upload
|
|
this._log.trace("Reconcile step 1");
|
|
if (item.id in this._tracker.changedIDs) {
|
|
if (this._isEqual(item))
|
|
this._tracker.removeChangedID(item.id);
|
|
self.done(false);
|
|
return;
|
|
}
|
|
|
|
// Step 2: Check for updates
|
|
// If different from local record, apply server update
|
|
this._log.trace("Reconcile step 2");
|
|
if (this._store.itemExists(item.id)) {
|
|
self.done(!this._isEqual(item));
|
|
return;
|
|
}
|
|
|
|
// If the incoming item has been deleted, skip step 3
|
|
this._log.trace("Reconcile step 2.5");
|
|
if (item.deleted) {
|
|
self.done(true);
|
|
return;
|
|
}
|
|
|
|
// Step 3: Check for similar items
|
|
this._log.trace("Reconcile step 3");
|
|
for (let id in this._tracker.changedIDs) {
|
|
let out = this._createRecord(id);
|
|
if (this._recordLike(item, out)) {
|
|
this._store.changeItemID(id, item.id);
|
|
this._tracker.removeChangedID(id);
|
|
this._tracker.removeChangedID(item.id);
|
|
this._store.cache.clear(); // because parentid refs will be wrong
|
|
self.done(false);
|
|
return;
|
|
}
|
|
}
|
|
self.done(true);
|
|
},
|
|
|
|
// Apply incoming records
|
|
_applyIncoming: function SyncEngine__applyIncoming(item) {
|
|
let self = yield;
|
|
this._log.trace("Incoming:\n" + item);
|
|
try {
|
|
this._tracker.ignoreAll = true;
|
|
yield this._store.applyIncoming(self.cb, item);
|
|
if (this._lastSyncTmp < item.modified)
|
|
this._lastSyncTmp = item.modified;
|
|
} catch (e) {
|
|
this._log.warn("Error while applying incoming record: " +
|
|
(e.message? e.message : e));
|
|
} finally {
|
|
this._tracker.ignoreAll = false;
|
|
}
|
|
},
|
|
|
|
// Upload outgoing records
|
|
_uploadOutgoing: function SyncEngine__uploadOutgoing() {
|
|
let self = yield;
|
|
|
|
let outnum = [i for (i in this._tracker.changedIDs)].length;
|
|
this._log.debug("Preparing " + outnum + " outgoing records");
|
|
if (outnum) {
|
|
// collection we'll upload
|
|
let up = new Collection(this.engineURL);
|
|
let meta = {};
|
|
|
|
// don't cache the outgoing items, we won't need them later
|
|
this._store.cache.enabled = false;
|
|
|
|
for (let id in this._tracker.changedIDs) {
|
|
let out = this._createRecord(id);
|
|
this._log.trace("Outgoing:\n" + out);
|
|
// skip getting siblings of already processed and deleted records
|
|
if (!out.deleted && !(out.id in meta))
|
|
this._store.createMetaRecords(out.id, meta);
|
|
yield out.encrypt(self.cb, ID.get('WeaveCryptoID').password);
|
|
up.pushData(JSON.parse(out.serialize())); // FIXME: inefficient
|
|
}
|
|
|
|
this._store.cache.enabled = true;
|
|
|
|
// now add short depth-and-index-only records, except the ones we're
|
|
// sending as full records
|
|
let count = 0;
|
|
for each (let obj in meta) {
|
|
if (!(obj.id in this._tracker.changedIDs)) {
|
|
up.pushData(obj);
|
|
count++;
|
|
}
|
|
}
|
|
|
|
this._log.info("Uploading " + outnum + " records + " + count + " index/depth records)");
|
|
// do the upload
|
|
yield up.post(self.cb);
|
|
|
|
// save last modified date
|
|
let mod = up.data.modified;
|
|
if (mod > this.lastSync)
|
|
this.lastSync = mod;
|
|
}
|
|
this._tracker.clearChangedIDs();
|
|
},
|
|
|
|
// Any cleanup necessary.
|
|
// Save the current snapshot so as to calculate changes at next sync
|
|
_syncFinish: function SyncEngine__syncFinish(error) {
|
|
let self = yield;
|
|
this._log.debug("Finishing up sync");
|
|
this._tracker.resetScore();
|
|
},
|
|
|
|
_sync: function SyncEngine__sync() {
|
|
let self = yield;
|
|
|
|
try {
|
|
yield this._syncStartup.async(this, self.cb);
|
|
Observers.notify("weave:engine:sync:status", "process-incoming");
|
|
yield this._processIncoming.async(this, self.cb);
|
|
Observers.notify("weave:engine:sync:status", "upload-outgoing");
|
|
yield this._uploadOutgoing.async(this, self.cb);
|
|
yield this._syncFinish.async(this, self.cb);
|
|
}
|
|
catch (e) {
|
|
this._log.warn("Sync failed");
|
|
throw e;
|
|
}
|
|
},
|
|
|
|
_wipeServer: function SyncEngine__wipeServer() {
|
|
let self = yield;
|
|
let all = new Resource(this.engineURL);
|
|
yield all.delete(self.cb);
|
|
let crypto = new Resource(this.cryptoMetaURL);
|
|
yield crypto.delete(self.cb);
|
|
},
|
|
|
|
_resetClient: function SyncEngine__resetClient() {
|
|
let self = yield;
|
|
this.resetLastSync();
|
|
}
|
|
};
|