mirror of
https://github.com/Foundry376/Mailspring.git
synced 2025-01-10 18:23:21 +08:00
5691fc88ca
Summary: Previously we were storing sync cursors in config.cson. They were by far the most frequently updated piece of data in config. To make things worse, all the writing was happening in the worker window - the main window was just seeing the changes on disk and reloading. I believe there's an edge case which causes the main window to read the config file when it's mid-write and empty. This causes the accounts array to become empty in the AccountStore and lots of downstream issues. It's also then possible for the main window to write a config change of it's own and empty the file permanently. Test Plan: A few new tests to make sure this is backwards compatible. Reviewers: juan, evan Reviewed By: juan, evan Differential Revision: https://phab.nylas.com/D2642
232 lines
6.3 KiB
CoffeeScript
232 lines
6.3 KiB
CoffeeScript
_ = require 'underscore'
|
|
{Actions, DatabaseStore} = require 'nylas-exports'
|
|
NylasLongConnection = require './nylas-long-connection'
|
|
ContactRankingsCache = require './contact-rankings-cache'
|
|
|
|
INITIAL_PAGE_SIZE = 30
|
|
MAX_PAGE_SIZE = 200
|
|
|
|
# BackoffTimer is a small helper class that wraps setTimeout. It fires the function
|
|
# you provide at a regular interval, but backs off each time you call `backoff`.
|
|
#
|
|
class BackoffTimer
|
|
constructor: (@fn) ->
|
|
@reset()
|
|
|
|
cancel: =>
|
|
clearTimeout(@_timeout) if @_timeout
|
|
@_timeout = null
|
|
|
|
reset: =>
|
|
@cancel()
|
|
@_delay = 20 * 1000
|
|
|
|
backoff: =>
|
|
@_delay = Math.min(@_delay * 1.4, 5 * 1000 * 60) # Cap at 5 minutes
|
|
if not NylasEnv.inSpecMode()
|
|
console.log("Backing off after sync failure. Will retry in #{Math.floor(@_delay / 1000)} seconds.")
|
|
|
|
start: =>
|
|
clearTimeout(@_timeout) if @_timeout
|
|
@_timeout = setTimeout =>
|
|
@_timeout = null
|
|
@fn()
|
|
, @_delay
|
|
|
|
|
|
module.exports =
|
|
class NylasSyncWorker
|
|
|
|
constructor: (api, account) ->
|
|
@_api = api
|
|
@_account = account
|
|
|
|
@_terminated = false
|
|
@_connection = new NylasLongConnection(api, account.id, {
|
|
ready: => @_state isnt null
|
|
getCursor: =>
|
|
return null if @_state is null
|
|
@_state['cursor'] || NylasEnv.config.get("nylas.#{@_account.id}.cursor")
|
|
setCursor: (val) =>
|
|
@_state['cursor'] = val
|
|
@writeState()
|
|
})
|
|
|
|
@_refreshingCaches = [new ContactRankingsCache(account.id)]
|
|
@_resumeTimer = new BackoffTimer =>
|
|
# indirection needed so resumeFetches can be spied on
|
|
@resumeFetches()
|
|
|
|
@_unlisten = Actions.retryInitialSync.listen(@_onRetryInitialSync, @)
|
|
|
|
@_state = null
|
|
DatabaseStore.findJSONBlob("NylasSyncWorker:#{@_account.id}").then (json) =>
|
|
@_state = json ? {}
|
|
for key in ['threads', 'labels', 'folders', 'drafts', 'contacts', 'calendars', 'events']
|
|
@_state[key].busy = false if @_state[key]
|
|
@resumeFetches()
|
|
@_connection.start()
|
|
|
|
@
|
|
|
|
account: ->
|
|
@_account
|
|
|
|
connection: ->
|
|
@_connection
|
|
|
|
state: ->
|
|
@_state
|
|
|
|
busy: ->
|
|
return false unless @_state
|
|
for key, state of @_state
|
|
if state.busy
|
|
return true
|
|
false
|
|
|
|
start: ->
|
|
@_resumeTimer.start()
|
|
@_connection.start()
|
|
@_refreshingCaches.map (c) -> c.start()
|
|
@resumeFetches()
|
|
|
|
cleanup: ->
|
|
@_unlisten?()
|
|
@_resumeTimer.cancel()
|
|
@_connection.end()
|
|
@_refreshingCaches.map (c) -> c.end()
|
|
@_terminated = true
|
|
@
|
|
|
|
resumeFetches: =>
|
|
return unless @_state
|
|
|
|
# Stop the timer. If one or more network requests fails during the fetch process
|
|
# we'll backoff and restart the timer.
|
|
@_resumeTimer.cancel()
|
|
|
|
@fetchCollection('threads')
|
|
if @_account.usesLabels()
|
|
@fetchCollection('labels', {initialPageSize: 1000})
|
|
if @_account.usesFolders()
|
|
@fetchCollection('folders', {initialPageSize: 1000})
|
|
@fetchCollection('drafts')
|
|
@fetchCollection('contacts')
|
|
@fetchCollection('calendars')
|
|
@fetchCollection('events')
|
|
|
|
fetchCollection: (model, options = {}) ->
|
|
return unless @_state
|
|
state = @_state[model] ? {}
|
|
|
|
return if state.complete and not options.force?
|
|
return if state.busy
|
|
|
|
state.complete = false
|
|
state.error = null
|
|
state.busy = true
|
|
state.fetched ?= 0
|
|
|
|
if not state.count
|
|
state.count = 0
|
|
@fetchCollectionCount(model)
|
|
|
|
if state.errorRequestRange
|
|
{limit, offset} = state.errorRequestRange
|
|
state.errorRequestRange = null
|
|
@fetchCollectionPage(model, {limit, offset})
|
|
else
|
|
@fetchCollectionPage(model, {
|
|
limit: options.initialPageSize ? INITIAL_PAGE_SIZE,
|
|
offset: 0
|
|
})
|
|
|
|
@_state[model] = state
|
|
@writeState()
|
|
|
|
fetchCollectionCount: (model) ->
|
|
@_api.makeRequest
|
|
accountId: @_account.id
|
|
path: "/#{model}"
|
|
returnsModel: false
|
|
qs:
|
|
view: 'count'
|
|
success: (response) =>
|
|
return if @_terminated
|
|
@updateTransferState(model, count: response.count)
|
|
error: (err) =>
|
|
return if @_terminated
|
|
@_resumeTimer.backoff()
|
|
@_resumeTimer.start()
|
|
|
|
fetchCollectionPage: (model, params = {}) ->
|
|
requestStartTime = Date.now()
|
|
requestOptions =
|
|
error: (err) =>
|
|
return if @_terminated
|
|
@_fetchCollectionPageError(model, params, err)
|
|
|
|
success: (json) =>
|
|
return if @_terminated
|
|
|
|
if model in ["labels", "folders"] and @_hasNoInbox(json)
|
|
@_fetchCollectionPageError(model, params, "No inbox in #{model}")
|
|
return
|
|
|
|
lastReceivedIndex = params.offset + json.length
|
|
moreToFetch = json.length is params.limit
|
|
|
|
if moreToFetch
|
|
nextParams = _.extend({}, params, {offset: lastReceivedIndex})
|
|
nextParams.limit = Math.min(Math.round(params.limit * 1.5), MAX_PAGE_SIZE)
|
|
nextDelay = Math.max(0, 1000 - (Date.now() - requestStartTime))
|
|
setTimeout(( => @fetchCollectionPage(model, nextParams)), nextDelay)
|
|
|
|
@updateTransferState(model, {
|
|
fetched: lastReceivedIndex,
|
|
busy: moreToFetch,
|
|
complete: !moreToFetch,
|
|
error: null,
|
|
errorRequestRange: null
|
|
})
|
|
|
|
if model is 'threads'
|
|
@_api.getThreads(@_account.id, params, requestOptions)
|
|
else
|
|
@_api.getCollection(@_account.id, model, params, requestOptions)
|
|
|
|
# It's occasionally possible for the NylasAPI's labels or folders
|
|
# endpoint to not return an "inbox" label. Since that's a core part of
|
|
# the app and it doesn't function without it, keep retrying until we see
|
|
# it.
|
|
_hasNoInbox: (json) ->
|
|
return not _.any(json, (obj) -> obj.name is "inbox")
|
|
|
|
_fetchCollectionPageError: (model, params, err) ->
|
|
@_resumeTimer.backoff()
|
|
@_resumeTimer.start()
|
|
@updateTransferState(model, {
|
|
busy: false,
|
|
complete: false,
|
|
error: err.toString()
|
|
errorRequestRange: {offset: params.offset, limit: params.limit}
|
|
})
|
|
|
|
updateTransferState: (model, updatedKeys) ->
|
|
@_state[model] = _.extend(@_state[model], updatedKeys)
|
|
@writeState()
|
|
|
|
writeState: ->
|
|
@_writeState ?= _.debounce =>
|
|
DatabaseStore.inTransaction (t) =>
|
|
t.persistJSONBlob("NylasSyncWorker:#{@_account.id}", @_state)
|
|
,100
|
|
@_writeState()
|
|
|
|
_onRetryInitialSync: =>
|
|
@resumeFetches()
|
|
|
|
NylasSyncWorker.BackoffTimer = BackoffTimer
|
|
NylasSyncWorker.INITIAL_PAGE_SIZE = INITIAL_PAGE_SIZE
|
|
NylasSyncWorker.MAX_PAGE_SIZE = MAX_PAGE_SIZE
|