refactor: cancel/done arch

Co-authored-by: Tsvetan Dimitrov tsvetan.dimitrov@avalabs.org
This commit is contained in:
Jonathan Oppenheimer 2026-01-12 12:07:18 -05:00
parent 6d048460ce
commit 75905dd04b
No known key found for this signature in database
GPG key ID: E4CEF9010EB8B740
5 changed files with 139 additions and 161 deletions

View file

@ -19,7 +19,6 @@ package snapshot
import ( import (
"bytes" "bytes"
"sync" "sync"
"sync/atomic"
"time" "time"
"github.com/VictoriaMetrics/fastcache" "github.com/VictoriaMetrics/fastcache"
@ -27,7 +26,6 @@ import (
"github.com/ethereum/go-ethereum/core/rawdb" "github.com/ethereum/go-ethereum/core/rawdb"
"github.com/ethereum/go-ethereum/core/types" "github.com/ethereum/go-ethereum/core/types"
"github.com/ethereum/go-ethereum/ethdb" "github.com/ethereum/go-ethereum/ethdb"
"github.com/ethereum/go-ethereum/log"
"github.com/ethereum/go-ethereum/rlp" "github.com/ethereum/go-ethereum/rlp"
"github.com/ethereum/go-ethereum/triedb" "github.com/ethereum/go-ethereum/triedb"
) )
@ -41,10 +39,18 @@ type diskLayer struct {
root common.Hash // Root hash of the base snapshot root common.Hash // Root hash of the base snapshot
stale bool // Signals that the layer became stale (state progressed) stale bool // Signals that the layer became stale (state progressed)
genMarker []byte // Marker for the state that's indexed during initial layer generation genMarker []byte // Marker for the state that's indexed during initial layer generation
genPending chan struct{} // Notification channel when generation is done (test synchronicity) genPending chan struct{} // Notification channel when generation is done (test synchronicity)
genAbort chan chan *generatorStats // Notification channel to abort generating the snapshot in this layer
genRunning atomic.Bool // Tracks whether the generator goroutine is actually running // Generator lifecycle management:
// - [cancel] is closed to request termination (broadcast).
// - [done] is closed by the generator goroutine on exit.
cancel chan struct{}
done chan struct{}
cancelOnce sync.Once
genStats *generatorStats // Stats for snapshot generation (generation aborted/finished if non-nil)
abortStarted time.Time // Timestamp when abort was first requested
lock sync.RWMutex lock sync.RWMutex
} }
@ -192,36 +198,35 @@ func (dl *diskLayer) Update(blockHash common.Hash, accounts map[common.Hash][]by
return newDiffLayer(dl, blockHash, accounts, storage) return newDiffLayer(dl, blockHash, accounts, storage)
} }
// stopGeneration aborts the state snapshot generation if it is currently running. // stopGeneration requests cancellation of any running snapshot generation and
// blocks until the generator goroutine (if running) has fully terminated.
//
// Concurrency guarantees:
// - Thread-safe: May be called concurrently from multiple goroutines
// - Idempotent: Safe to call multiple times; subsequent calls have no effect
// - Blocking: Returns only after the generator goroutine (if any) has exited
// - Safe to call at any time, including when no generation is running
//
// After return, it is **guaranteed** that:
// - The generator goroutine has terminated
// - It is safe to proceed with cleanup operations (e.g. closing databases)
func (dl *diskLayer) stopGeneration() { func (dl *diskLayer) stopGeneration() {
// Check if generation goroutine is actually running cancel := dl.cancel
done := dl.done
if cancel == nil || done == nil {
return
}
// Store ideal time for abort to get better estimate of load
// //
// Note: genMarker can be nil even when the generator is still running (waiting // Note that we set this time regardless if abortion was skipped otherwise we
// for abort signal after completing generation), so we can't rely on genMarker. // will never restart generation (age will always be negative).
if !dl.genRunning.Load() { if dl.abortStarted.IsZero() {
return dl.abortStarted = time.Now()
} }
// Use write lock to ensure only one goroutine can stop generation at a time, dl.cancelOnce.Do(func() {
// preventing a race where multiple callers might try to send abort signals. close(cancel)
dl.lock.Lock() })
genAbort := dl.genAbort <-done
if genAbort == nil {
dl.lock.Unlock()
return
}
// Clear genAbort immediately while holding the lock to prevent other callers
// from attempting to use the same channel
dl.genAbort = nil
dl.lock.Unlock()
// Perform the channel handshake without holding the lock to avoid deadlocks.
abort := make(chan *generatorStats)
select {
case genAbort <- abort:
// Generator received the abort signal, wait for it to respond
<-abort
case <-time.After(5 * time.Second):
log.Error("Snapshot generator did not respond despite being marked as running")
}
} }

View file

@ -50,6 +50,9 @@ var (
// errMissingTrie is returned if the target trie is missing while the generation // errMissingTrie is returned if the target trie is missing while the generation
// is running. In this case the generation is aborted and wait the new signal. // is running. In this case the generation is aborted and wait the new signal.
errMissingTrie = errors.New("missing trie") errMissingTrie = errors.New("missing trie")
// errAborted is returned when snapshot generation was interrupted/aborted
errAborted = errors.New("aborted")
) )
// generateSnapshot regenerates a brand new snapshot based on an existing state // generateSnapshot regenerates a brand new snapshot based on an existing state
@ -74,7 +77,8 @@ func generateSnapshot(diskdb ethdb.KeyValueStore, triedb *triedb.Database, cache
cache: fastcache.New(cache * 1024 * 1024), cache: fastcache.New(cache * 1024 * 1024),
genMarker: genMarker, genMarker: genMarker,
genPending: make(chan struct{}), genPending: make(chan struct{}),
genAbort: make(chan chan *generatorStats), cancel: make(chan struct{}),
done: make(chan struct{}),
} }
go base.generate(stats) go base.generate(stats)
log.Debug("Start snapshot generation", "root", root) log.Debug("Start snapshot generation", "root", root)
@ -467,12 +471,14 @@ func (dl *diskLayer) generateRange(ctx *generatorContext, trieId *trie.ID, prefi
// checkAndFlush checks if an interruption signal is received or the // checkAndFlush checks if an interruption signal is received or the
// batch size has exceeded the allowance. // batch size has exceeded the allowance.
func (dl *diskLayer) checkAndFlush(ctx *generatorContext, current []byte) error { func (dl *diskLayer) checkAndFlush(ctx *generatorContext, current []byte) error {
var abort chan *generatorStats aborting := false
select { select {
case abort = <-dl.genAbort: case <-dl.cancel:
aborting = true
default: default:
} }
if ctx.batch.ValueSize() > ethdb.IdealBatchSize || abort != nil {
if ctx.batch.ValueSize() > ethdb.IdealBatchSize || aborting {
if bytes.Compare(current, dl.genMarker) < 0 { if bytes.Compare(current, dl.genMarker) < 0 {
log.Error("Snapshot generator went backwards", "current", fmt.Sprintf("%x", current), "genMarker", fmt.Sprintf("%x", dl.genMarker)) log.Error("Snapshot generator went backwards", "current", fmt.Sprintf("%x", current), "genMarker", fmt.Sprintf("%x", dl.genMarker))
} }
@ -482,6 +488,7 @@ func (dl *diskLayer) checkAndFlush(ctx *generatorContext, current []byte) error
journalProgress(ctx.batch, current, ctx.stats) journalProgress(ctx.batch, current, ctx.stats)
if err := ctx.batch.Write(); err != nil { if err := ctx.batch.Write(); err != nil {
log.Error("Failed to flush batch", "err", err)
return err return err
} }
ctx.batch.Reset() ctx.batch.Reset()
@ -490,9 +497,9 @@ func (dl *diskLayer) checkAndFlush(ctx *generatorContext, current []byte) error
dl.genMarker = current dl.genMarker = current
dl.lock.Unlock() dl.lock.Unlock()
if abort != nil { if aborting {
ctx.stats.Log("Aborting state snapshot generation", dl.root, current) ctx.stats.Log("Aborting state snapshot generation", dl.root, current)
return newAbortErr(abort) // bubble up an error for interruption return errAborted
} }
// Don't hold the iterators too long, release them to let compactor works // Don't hold the iterators too long, release them to let compactor works
ctx.reopenIterator(snapAccount) ctx.reopenIterator(snapAccount)
@ -648,13 +655,11 @@ func generateAccounts(ctx *generatorContext, dl *diskLayer, accMarker []byte) er
// gathering and logging, since the method surfs the blocks as they arrive, often // gathering and logging, since the method surfs the blocks as they arrive, often
// being restarted. // being restarted.
func (dl *diskLayer) generate(stats *generatorStats) { func (dl *diskLayer) generate(stats *generatorStats) {
dl.genRunning.Store(true) if dl.done != nil {
defer dl.genRunning.Store(false) defer close(dl.done)
}
var ( var accMarker []byte
accMarker []byte
abort chan *generatorStats
)
if len(dl.genMarker) > 0 { // []byte{} is the start, use nil for that if len(dl.genMarker) > 0 { // []byte{} is the start, use nil for that
accMarker = dl.genMarker[:common.HashLength] accMarker = dl.genMarker[:common.HashLength]
} }
@ -672,15 +677,11 @@ func (dl *diskLayer) generate(stats *generatorStats) {
defer ctx.close() defer ctx.close()
if err := generateAccounts(ctx, dl, accMarker); err != nil { if err := generateAccounts(ctx, dl, accMarker); err != nil {
// Extract the received interruption signal if exists // Check if error was due to abort
if aerr, ok := err.(*abortErr); ok { if err == errAborted {
abort = aerr.abort stats.Log("Aborting state snapshot generation", dl.root, dl.genMarker)
} }
// Aborted by internal error, wait the signal dl.genStats = stats
if abort == nil {
abort = <-dl.genAbort
}
abort <- stats
return return
} }
// Snapshot fully generated, set the marker to nil. // Snapshot fully generated, set the marker to nil.
@ -689,9 +690,7 @@ func (dl *diskLayer) generate(stats *generatorStats) {
journalProgress(ctx.batch, nil, stats) journalProgress(ctx.batch, nil, stats)
if err := ctx.batch.Write(); err != nil { if err := ctx.batch.Write(); err != nil {
log.Error("Failed to flush batch", "err", err) log.Error("Failed to flush batch", "err", err)
dl.genStats = stats
abort = <-dl.genAbort
abort <- stats
return return
} }
ctx.batch.Reset() ctx.batch.Reset()
@ -701,12 +700,9 @@ func (dl *diskLayer) generate(stats *generatorStats) {
dl.lock.Lock() dl.lock.Lock()
dl.genMarker = nil dl.genMarker = nil
dl.genStats = stats
close(dl.genPending) close(dl.genPending)
dl.lock.Unlock() dl.lock.Unlock()
// Someone will be looking for us, wait it out
abort = <-dl.genAbort
abort <- nil
} }
// increaseKey increase the input key by one bit. Return nil if the entire // increaseKey increase the input key by one bit. Return nil if the entire
@ -720,17 +716,3 @@ func increaseKey(key []byte) []byte {
} }
return nil return nil
} }
// abortErr wraps an interruption signal received to represent the
// generation is aborted by external processes.
type abortErr struct {
abort chan *generatorStats
}
func newAbortErr(abort chan *generatorStats) error {
return &abortErr{abort: abort}
}
func (err *abortErr) Error() string {
return "aborted"
}

View file

@ -79,10 +79,10 @@ func testGeneration(t *testing.T, scheme string) {
} }
checkSnapRoot(t, snap, root) checkSnapRoot(t, snap, root)
// Signal abortion to the generator and wait for it to tear down // Stop the generator (if still running) and wait for it to exit.
stop := make(chan *generatorStats) if err := snap.Release(); err != nil {
snap.genAbort <- stop t.Fatal(err)
<-stop }
} }
// Tests that snapshot generation with existent flat state. // Tests that snapshot generation with existent flat state.
@ -120,10 +120,10 @@ func testGenerateExistentState(t *testing.T, scheme string) {
} }
checkSnapRoot(t, snap, root) checkSnapRoot(t, snap, root)
// Signal abortion to the generator and wait for it to tear down // Stop the generator (if still running) and wait for it to exit.
stop := make(chan *generatorStats) if err := snap.Release(); err != nil {
snap.genAbort <- stop t.Fatal(err)
<-stop }
} }
func checkSnapRoot(t *testing.T, snap *diskLayer, trieRoot common.Hash) { func checkSnapRoot(t *testing.T, snap *diskLayer, trieRoot common.Hash) {
@ -356,10 +356,10 @@ func testGenerateExistentStateWithWrongStorage(t *testing.T, scheme string) {
t.Errorf("Snapshot generation failed") t.Errorf("Snapshot generation failed")
} }
checkSnapRoot(t, snap, root) checkSnapRoot(t, snap, root)
// Signal abortion to the generator and wait for it to tear down // Stop the generator (if still running) and wait for it to exit.
stop := make(chan *generatorStats) if err := snap.Release(); err != nil {
snap.genAbort <- stop t.Fatal(err)
<-stop }
} }
// Tests that snapshot generation with existent flat state, where the flat state // Tests that snapshot generation with existent flat state, where the flat state
@ -419,10 +419,10 @@ func testGenerateExistentStateWithWrongAccounts(t *testing.T, scheme string) {
} }
checkSnapRoot(t, snap, root) checkSnapRoot(t, snap, root)
// Signal abortion to the generator and wait for it to tear down // Stop the generator (if still running) and wait for it to exit.
stop := make(chan *generatorStats) if err := snap.Release(); err != nil {
snap.genAbort <- stop t.Fatal(err)
<-stop }
} }
// Tests that snapshot generation errors out correctly in case of a missing trie // Tests that snapshot generation errors out correctly in case of a missing trie
@ -459,10 +459,10 @@ func testGenerateCorruptAccountTrie(t *testing.T, scheme string) {
case <-time.After(time.Second): case <-time.After(time.Second):
// Not generated fast enough, hopefully blocked inside on missing trie node fail // Not generated fast enough, hopefully blocked inside on missing trie node fail
} }
// Signal abortion to the generator and wait for it to tear down // Stop the generator (if still running) and wait for it to exit.
stop := make(chan *generatorStats) if err := snap.Release(); err != nil {
snap.genAbort <- stop t.Fatal(err)
<-stop }
} }
// Tests that snapshot generation errors out correctly in case of a missing root // Tests that snapshot generation errors out correctly in case of a missing root
@ -503,10 +503,10 @@ func testGenerateMissingStorageTrie(t *testing.T, scheme string) {
case <-time.After(time.Second): case <-time.After(time.Second):
// Not generated fast enough, hopefully blocked inside on missing trie node fail // Not generated fast enough, hopefully blocked inside on missing trie node fail
} }
// Signal abortion to the generator and wait for it to tear down // Stop the generator (if still running) and wait for it to exit.
stop := make(chan *generatorStats) if err := snap.Release(); err != nil {
snap.genAbort <- stop t.Fatal(err)
<-stop }
} }
// Tests that snapshot generation errors out correctly in case of a missing trie // Tests that snapshot generation errors out correctly in case of a missing trie
@ -545,10 +545,10 @@ func testGenerateCorruptStorageTrie(t *testing.T, scheme string) {
case <-time.After(time.Second): case <-time.After(time.Second):
// Not generated fast enough, hopefully blocked inside on missing trie node fail // Not generated fast enough, hopefully blocked inside on missing trie node fail
} }
// Signal abortion to the generator and wait for it to tear down // Stop the generator (if still running) and wait for it to exit.
stop := make(chan *generatorStats) if err := snap.Release(); err != nil {
snap.genAbort <- stop t.Fatal(err)
<-stop }
} }
// Tests that snapshot generation when an extra account with storage exists in the snap state. // Tests that snapshot generation when an extra account with storage exists in the snap state.
@ -610,10 +610,10 @@ func testGenerateWithExtraAccounts(t *testing.T, scheme string) {
} }
checkSnapRoot(t, snap, root) checkSnapRoot(t, snap, root)
// Signal abortion to the generator and wait for it to tear down // Stop the generator (if still running) and wait for it to exit.
stop := make(chan *generatorStats) if err := snap.Release(); err != nil {
snap.genAbort <- stop t.Fatal(err)
<-stop }
// If we now inspect the snap db, there should exist no extraneous storage items // If we now inspect the snap db, there should exist no extraneous storage items
if data := rawdb.ReadStorageSnapshot(helper.diskdb, hashData([]byte("acc-2")), hashData([]byte("b-key-1"))); data != nil { if data := rawdb.ReadStorageSnapshot(helper.diskdb, hashData([]byte("acc-2")), hashData([]byte("b-key-1"))); data != nil {
t.Fatalf("expected slot to be removed, got %v", string(data)) t.Fatalf("expected slot to be removed, got %v", string(data))
@ -671,10 +671,10 @@ func testGenerateWithManyExtraAccounts(t *testing.T, scheme string) {
t.Errorf("Snapshot generation failed") t.Errorf("Snapshot generation failed")
} }
checkSnapRoot(t, snap, root) checkSnapRoot(t, snap, root)
// Signal abortion to the generator and wait for it to tear down // Stop the generator (if still running) and wait for it to exit.
stop := make(chan *generatorStats) if err := snap.Release(); err != nil {
snap.genAbort <- stop t.Fatal(err)
<-stop }
} }
// Tests this case // Tests this case
@ -720,10 +720,10 @@ func testGenerateWithExtraBeforeAndAfter(t *testing.T, scheme string) {
t.Errorf("Snapshot generation failed") t.Errorf("Snapshot generation failed")
} }
checkSnapRoot(t, snap, root) checkSnapRoot(t, snap, root)
// Signal abortion to the generator and wait for it to tear down // Stop the generator (if still running) and wait for it to exit.
stop := make(chan *generatorStats) if err := snap.Release(); err != nil {
snap.genAbort <- stop t.Fatal(err)
<-stop }
} }
// TestGenerateWithMalformedSnapdata tests what happes if we have some junk // TestGenerateWithMalformedSnapdata tests what happes if we have some junk
@ -760,10 +760,10 @@ func testGenerateWithMalformedSnapdata(t *testing.T, scheme string) {
t.Errorf("Snapshot generation failed") t.Errorf("Snapshot generation failed")
} }
checkSnapRoot(t, snap, root) checkSnapRoot(t, snap, root)
// Signal abortion to the generator and wait for it to tear down // Stop the generator (if still running) and wait for it to exit.
stop := make(chan *generatorStats) if err := snap.Release(); err != nil {
snap.genAbort <- stop t.Fatal(err)
<-stop }
// If we now inspect the snap db, there should exist no extraneous storage items // If we now inspect the snap db, there should exist no extraneous storage items
if data := rawdb.ReadStorageSnapshot(helper.diskdb, hashData([]byte("acc-2")), hashData([]byte("b-key-1"))); data != nil { if data := rawdb.ReadStorageSnapshot(helper.diskdb, hashData([]byte("acc-2")), hashData([]byte("b-key-1"))); data != nil {
t.Fatalf("expected slot to be removed, got %v", string(data)) t.Fatalf("expected slot to be removed, got %v", string(data))
@ -797,10 +797,10 @@ func testGenerateFromEmptySnap(t *testing.T, scheme string) {
t.Errorf("Snapshot generation failed") t.Errorf("Snapshot generation failed")
} }
checkSnapRoot(t, snap, root) checkSnapRoot(t, snap, root)
// Signal abortion to the generator and wait for it to tear down // Stop the generator (if still running) and wait for it to exit.
stop := make(chan *generatorStats) if err := snap.Release(); err != nil {
snap.genAbort <- stop t.Fatal(err)
<-stop }
} }
// Tests that snapshot generation with existent flat state, where the flat state // Tests that snapshot generation with existent flat state, where the flat state
@ -848,10 +848,10 @@ func testGenerateWithIncompleteStorage(t *testing.T, scheme string) {
t.Errorf("Snapshot generation failed") t.Errorf("Snapshot generation failed")
} }
checkSnapRoot(t, snap, root) checkSnapRoot(t, snap, root)
// Signal abortion to the generator and wait for it to tear down // Stop the generator (if still running) and wait for it to exit.
stop := make(chan *generatorStats) if err := snap.Release(); err != nil {
snap.genAbort <- stop t.Fatal(err)
<-stop }
} }
func incKey(key []byte) []byte { func incKey(key []byte) []byte {
@ -944,10 +944,10 @@ func testGenerateCompleteSnapshotWithDanglingStorage(t *testing.T, scheme string
} }
checkSnapRoot(t, snap, root) checkSnapRoot(t, snap, root)
// Signal abortion to the generator and wait for it to tear down // Stop the generator (if still running) and wait for it to exit.
stop := make(chan *generatorStats) if err := snap.Release(); err != nil {
snap.genAbort <- stop t.Fatal(err)
<-stop }
} }
// Tests that snapshot generation with dangling storages. Dangling storage means // Tests that snapshot generation with dangling storages. Dangling storage means
@ -981,10 +981,10 @@ func testGenerateBrokenSnapshotWithDanglingStorage(t *testing.T, scheme string)
} }
checkSnapRoot(t, snap, root) checkSnapRoot(t, snap, root)
// Signal abortion to the generator and wait for it to tear down // Stop the generator (if still running) and wait for it to exit.
stop := make(chan *generatorStats) if err := snap.Release(); err != nil {
snap.genAbort <- stop t.Fatal(err)
<-stop }
} }
// TestReleaseStopsGeneration verifies that Release() properly stops ongoing // TestReleaseStopsGeneration verifies that Release() properly stops ongoing

View file

@ -179,7 +179,8 @@ func loadSnapshot(diskdb ethdb.KeyValueStore, triedb *triedb.Database, root comm
// if the background generation is allowed // if the background generation is allowed
if !generator.Done && !noBuild { if !generator.Done && !noBuild {
base.genPending = make(chan struct{}) base.genPending = make(chan struct{})
base.genAbort = make(chan chan *generatorStats) base.cancel = make(chan struct{})
base.done = make(chan struct{})
var origin uint64 var origin uint64
if len(generator.Marker) >= 8 { if len(generator.Marker) >= 8 {
@ -199,16 +200,9 @@ func loadSnapshot(diskdb ethdb.KeyValueStore, triedb *triedb.Database, root comm
// Journal terminates any in-progress snapshot generation, also implicitly pushing // Journal terminates any in-progress snapshot generation, also implicitly pushing
// the progress into the database. // the progress into the database.
func (dl *diskLayer) Journal(buffer *bytes.Buffer) (common.Hash, error) { func (dl *diskLayer) Journal(buffer *bytes.Buffer) (common.Hash, error) {
// If the snapshot is currently being generated, abort it // If the snapshot is currently being generated, stop it
var stats *generatorStats dl.stopGeneration()
if dl.genAbort != nil {
abort := make(chan *generatorStats)
dl.genAbort <- abort
if stats = <-abort; stats != nil {
stats.Log("Journalling in-progress snapshot", dl.root, dl.genMarker)
}
}
// Ensure the layer didn't get stale // Ensure the layer didn't get stale
dl.lock.RLock() dl.lock.RLock()
defer dl.lock.RUnlock() defer dl.lock.RUnlock()
@ -216,8 +210,8 @@ func (dl *diskLayer) Journal(buffer *bytes.Buffer) (common.Hash, error) {
if dl.stale { if dl.stale {
return common.Hash{}, ErrSnapshotStale return common.Hash{}, ErrSnapshotStale
} }
// Ensure the generator stats is written even if none was ran this cycle // Ensure the generator marker is written even if none was ran this cycle
journalProgress(dl.diskdb, dl.genMarker, stats) journalProgress(dl.diskdb, dl.genMarker, nil)
log.Debug("Journalled disk layer", "root", dl.root) log.Debug("Journalled disk layer", "root", dl.root)
return dl.root, nil return dl.root, nil

View file

@ -492,7 +492,7 @@ func (t *Tree) cap(diff *diffLayer, layers int) *diskLayer {
// there's a snapshot being generated currently. In that case, the trie // there's a snapshot being generated currently. In that case, the trie
// will move from underneath the generator so we **must** merge all the // will move from underneath the generator so we **must** merge all the
// partial data down into the snapshot and restart the generation. // partial data down into the snapshot and restart the generation.
if flattened.parent.(*diskLayer).genAbort == nil { if flattened.parent.(*diskLayer).cancel == nil {
return nil return nil
} }
} }
@ -520,14 +520,10 @@ func diffToDisk(bottom *diffLayer) *diskLayer {
var ( var (
base = bottom.parent.(*diskLayer) base = bottom.parent.(*diskLayer)
batch = base.diskdb.NewBatch() batch = base.diskdb.NewBatch()
stats *generatorStats
) )
// If the disk layer is running a snapshot generator, abort it // Attempt to stop generation (if not already stopped)
if base.genAbort != nil { base.stopGeneration()
abort := make(chan *generatorStats)
base.genAbort <- abort
stats = <-abort
}
// Put the deletion in the batch writer, flush all updates in the final step. // Put the deletion in the batch writer, flush all updates in the final step.
rawdb.DeleteSnapshotRoot(batch) rawdb.DeleteSnapshotRoot(batch)
@ -606,8 +602,8 @@ func diffToDisk(bottom *diffLayer) *diskLayer {
// Update the snapshot block marker and write any remainder data // Update the snapshot block marker and write any remainder data
rawdb.WriteSnapshotRoot(batch, bottom.root) rawdb.WriteSnapshotRoot(batch, bottom.root)
// Write out the generator progress marker and report // Write out the generator progress marker
journalProgress(batch, base.genMarker, stats) journalProgress(batch, base.genMarker, base.genStats)
// Flush all the updates in the single db operation. Ensure the // Flush all the updates in the single db operation. Ensure the
// disk layer transition is atomic. // disk layer transition is atomic.
@ -626,12 +622,13 @@ func diffToDisk(bottom *diffLayer) *diskLayer {
// If snapshot generation hasn't finished yet, port over all the starts and // If snapshot generation hasn't finished yet, port over all the starts and
// continue where the previous round left off. // continue where the previous round left off.
// //
// Note, the `base.genAbort` comparison is not used normally, it's checked // Note, the `base.genPending` comparison is not used normally, it's checked
// to allow the tests to play with the marker without triggering this path. // to allow the tests to play with the marker without triggering this path.
if base.genMarker != nil && base.genAbort != nil { if base.genMarker != nil && base.genPending != nil {
res.genMarker = base.genMarker res.genMarker = base.genMarker
res.genAbort = make(chan chan *generatorStats) res.cancel = make(chan struct{})
go res.generate(stats) res.done = make(chan struct{})
go res.generate(base.genStats)
} }
return res return res
} }