mirror of
https://github.com/ethereum/go-ethereum.git
synced 2026-05-14 12:06:40 +00:00
refactor: cancel/done arch
Co-authored-by: Tsvetan Dimitrov tsvetan.dimitrov@avalabs.org
This commit is contained in:
parent
6d048460ce
commit
75905dd04b
5 changed files with 139 additions and 161 deletions
|
|
@ -19,7 +19,6 @@ package snapshot
|
||||||
import (
|
import (
|
||||||
"bytes"
|
"bytes"
|
||||||
"sync"
|
"sync"
|
||||||
"sync/atomic"
|
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
"github.com/VictoriaMetrics/fastcache"
|
"github.com/VictoriaMetrics/fastcache"
|
||||||
|
|
@ -27,7 +26,6 @@ import (
|
||||||
"github.com/ethereum/go-ethereum/core/rawdb"
|
"github.com/ethereum/go-ethereum/core/rawdb"
|
||||||
"github.com/ethereum/go-ethereum/core/types"
|
"github.com/ethereum/go-ethereum/core/types"
|
||||||
"github.com/ethereum/go-ethereum/ethdb"
|
"github.com/ethereum/go-ethereum/ethdb"
|
||||||
"github.com/ethereum/go-ethereum/log"
|
|
||||||
"github.com/ethereum/go-ethereum/rlp"
|
"github.com/ethereum/go-ethereum/rlp"
|
||||||
"github.com/ethereum/go-ethereum/triedb"
|
"github.com/ethereum/go-ethereum/triedb"
|
||||||
)
|
)
|
||||||
|
|
@ -41,10 +39,18 @@ type diskLayer struct {
|
||||||
root common.Hash // Root hash of the base snapshot
|
root common.Hash // Root hash of the base snapshot
|
||||||
stale bool // Signals that the layer became stale (state progressed)
|
stale bool // Signals that the layer became stale (state progressed)
|
||||||
|
|
||||||
genMarker []byte // Marker for the state that's indexed during initial layer generation
|
genMarker []byte // Marker for the state that's indexed during initial layer generation
|
||||||
genPending chan struct{} // Notification channel when generation is done (test synchronicity)
|
genPending chan struct{} // Notification channel when generation is done (test synchronicity)
|
||||||
genAbort chan chan *generatorStats // Notification channel to abort generating the snapshot in this layer
|
|
||||||
genRunning atomic.Bool // Tracks whether the generator goroutine is actually running
|
// Generator lifecycle management:
|
||||||
|
// - [cancel] is closed to request termination (broadcast).
|
||||||
|
// - [done] is closed by the generator goroutine on exit.
|
||||||
|
cancel chan struct{}
|
||||||
|
done chan struct{}
|
||||||
|
cancelOnce sync.Once
|
||||||
|
|
||||||
|
genStats *generatorStats // Stats for snapshot generation (generation aborted/finished if non-nil)
|
||||||
|
abortStarted time.Time // Timestamp when abort was first requested
|
||||||
|
|
||||||
lock sync.RWMutex
|
lock sync.RWMutex
|
||||||
}
|
}
|
||||||
|
|
@ -192,36 +198,35 @@ func (dl *diskLayer) Update(blockHash common.Hash, accounts map[common.Hash][]by
|
||||||
return newDiffLayer(dl, blockHash, accounts, storage)
|
return newDiffLayer(dl, blockHash, accounts, storage)
|
||||||
}
|
}
|
||||||
|
|
||||||
// stopGeneration aborts the state snapshot generation if it is currently running.
|
// stopGeneration requests cancellation of any running snapshot generation and
|
||||||
|
// blocks until the generator goroutine (if running) has fully terminated.
|
||||||
|
//
|
||||||
|
// Concurrency guarantees:
|
||||||
|
// - Thread-safe: May be called concurrently from multiple goroutines
|
||||||
|
// - Idempotent: Safe to call multiple times; subsequent calls have no effect
|
||||||
|
// - Blocking: Returns only after the generator goroutine (if any) has exited
|
||||||
|
// - Safe to call at any time, including when no generation is running
|
||||||
|
//
|
||||||
|
// After return, it is **guaranteed** that:
|
||||||
|
// - The generator goroutine has terminated
|
||||||
|
// - It is safe to proceed with cleanup operations (e.g. closing databases)
|
||||||
func (dl *diskLayer) stopGeneration() {
|
func (dl *diskLayer) stopGeneration() {
|
||||||
// Check if generation goroutine is actually running
|
cancel := dl.cancel
|
||||||
|
done := dl.done
|
||||||
|
if cancel == nil || done == nil {
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
// Store ideal time for abort to get better estimate of load
|
||||||
//
|
//
|
||||||
// Note: genMarker can be nil even when the generator is still running (waiting
|
// Note that we set this time regardless if abortion was skipped otherwise we
|
||||||
// for abort signal after completing generation), so we can't rely on genMarker.
|
// will never restart generation (age will always be negative).
|
||||||
if !dl.genRunning.Load() {
|
if dl.abortStarted.IsZero() {
|
||||||
return
|
dl.abortStarted = time.Now()
|
||||||
}
|
}
|
||||||
|
|
||||||
// Use write lock to ensure only one goroutine can stop generation at a time,
|
dl.cancelOnce.Do(func() {
|
||||||
// preventing a race where multiple callers might try to send abort signals.
|
close(cancel)
|
||||||
dl.lock.Lock()
|
})
|
||||||
genAbort := dl.genAbort
|
<-done
|
||||||
if genAbort == nil {
|
|
||||||
dl.lock.Unlock()
|
|
||||||
return
|
|
||||||
}
|
|
||||||
// Clear genAbort immediately while holding the lock to prevent other callers
|
|
||||||
// from attempting to use the same channel
|
|
||||||
dl.genAbort = nil
|
|
||||||
dl.lock.Unlock()
|
|
||||||
|
|
||||||
// Perform the channel handshake without holding the lock to avoid deadlocks.
|
|
||||||
abort := make(chan *generatorStats)
|
|
||||||
select {
|
|
||||||
case genAbort <- abort:
|
|
||||||
// Generator received the abort signal, wait for it to respond
|
|
||||||
<-abort
|
|
||||||
case <-time.After(5 * time.Second):
|
|
||||||
log.Error("Snapshot generator did not respond despite being marked as running")
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
|
||||||
|
|
@ -50,6 +50,9 @@ var (
|
||||||
// errMissingTrie is returned if the target trie is missing while the generation
|
// errMissingTrie is returned if the target trie is missing while the generation
|
||||||
// is running. In this case the generation is aborted and wait the new signal.
|
// is running. In this case the generation is aborted and wait the new signal.
|
||||||
errMissingTrie = errors.New("missing trie")
|
errMissingTrie = errors.New("missing trie")
|
||||||
|
|
||||||
|
// errAborted is returned when snapshot generation was interrupted/aborted
|
||||||
|
errAborted = errors.New("aborted")
|
||||||
)
|
)
|
||||||
|
|
||||||
// generateSnapshot regenerates a brand new snapshot based on an existing state
|
// generateSnapshot regenerates a brand new snapshot based on an existing state
|
||||||
|
|
@ -74,7 +77,8 @@ func generateSnapshot(diskdb ethdb.KeyValueStore, triedb *triedb.Database, cache
|
||||||
cache: fastcache.New(cache * 1024 * 1024),
|
cache: fastcache.New(cache * 1024 * 1024),
|
||||||
genMarker: genMarker,
|
genMarker: genMarker,
|
||||||
genPending: make(chan struct{}),
|
genPending: make(chan struct{}),
|
||||||
genAbort: make(chan chan *generatorStats),
|
cancel: make(chan struct{}),
|
||||||
|
done: make(chan struct{}),
|
||||||
}
|
}
|
||||||
go base.generate(stats)
|
go base.generate(stats)
|
||||||
log.Debug("Start snapshot generation", "root", root)
|
log.Debug("Start snapshot generation", "root", root)
|
||||||
|
|
@ -467,12 +471,14 @@ func (dl *diskLayer) generateRange(ctx *generatorContext, trieId *trie.ID, prefi
|
||||||
// checkAndFlush checks if an interruption signal is received or the
|
// checkAndFlush checks if an interruption signal is received or the
|
||||||
// batch size has exceeded the allowance.
|
// batch size has exceeded the allowance.
|
||||||
func (dl *diskLayer) checkAndFlush(ctx *generatorContext, current []byte) error {
|
func (dl *diskLayer) checkAndFlush(ctx *generatorContext, current []byte) error {
|
||||||
var abort chan *generatorStats
|
aborting := false
|
||||||
select {
|
select {
|
||||||
case abort = <-dl.genAbort:
|
case <-dl.cancel:
|
||||||
|
aborting = true
|
||||||
default:
|
default:
|
||||||
}
|
}
|
||||||
if ctx.batch.ValueSize() > ethdb.IdealBatchSize || abort != nil {
|
|
||||||
|
if ctx.batch.ValueSize() > ethdb.IdealBatchSize || aborting {
|
||||||
if bytes.Compare(current, dl.genMarker) < 0 {
|
if bytes.Compare(current, dl.genMarker) < 0 {
|
||||||
log.Error("Snapshot generator went backwards", "current", fmt.Sprintf("%x", current), "genMarker", fmt.Sprintf("%x", dl.genMarker))
|
log.Error("Snapshot generator went backwards", "current", fmt.Sprintf("%x", current), "genMarker", fmt.Sprintf("%x", dl.genMarker))
|
||||||
}
|
}
|
||||||
|
|
@ -482,6 +488,7 @@ func (dl *diskLayer) checkAndFlush(ctx *generatorContext, current []byte) error
|
||||||
journalProgress(ctx.batch, current, ctx.stats)
|
journalProgress(ctx.batch, current, ctx.stats)
|
||||||
|
|
||||||
if err := ctx.batch.Write(); err != nil {
|
if err := ctx.batch.Write(); err != nil {
|
||||||
|
log.Error("Failed to flush batch", "err", err)
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
ctx.batch.Reset()
|
ctx.batch.Reset()
|
||||||
|
|
@ -490,9 +497,9 @@ func (dl *diskLayer) checkAndFlush(ctx *generatorContext, current []byte) error
|
||||||
dl.genMarker = current
|
dl.genMarker = current
|
||||||
dl.lock.Unlock()
|
dl.lock.Unlock()
|
||||||
|
|
||||||
if abort != nil {
|
if aborting {
|
||||||
ctx.stats.Log("Aborting state snapshot generation", dl.root, current)
|
ctx.stats.Log("Aborting state snapshot generation", dl.root, current)
|
||||||
return newAbortErr(abort) // bubble up an error for interruption
|
return errAborted
|
||||||
}
|
}
|
||||||
// Don't hold the iterators too long, release them to let compactor works
|
// Don't hold the iterators too long, release them to let compactor works
|
||||||
ctx.reopenIterator(snapAccount)
|
ctx.reopenIterator(snapAccount)
|
||||||
|
|
@ -648,13 +655,11 @@ func generateAccounts(ctx *generatorContext, dl *diskLayer, accMarker []byte) er
|
||||||
// gathering and logging, since the method surfs the blocks as they arrive, often
|
// gathering and logging, since the method surfs the blocks as they arrive, often
|
||||||
// being restarted.
|
// being restarted.
|
||||||
func (dl *diskLayer) generate(stats *generatorStats) {
|
func (dl *diskLayer) generate(stats *generatorStats) {
|
||||||
dl.genRunning.Store(true)
|
if dl.done != nil {
|
||||||
defer dl.genRunning.Store(false)
|
defer close(dl.done)
|
||||||
|
}
|
||||||
|
|
||||||
var (
|
var accMarker []byte
|
||||||
accMarker []byte
|
|
||||||
abort chan *generatorStats
|
|
||||||
)
|
|
||||||
if len(dl.genMarker) > 0 { // []byte{} is the start, use nil for that
|
if len(dl.genMarker) > 0 { // []byte{} is the start, use nil for that
|
||||||
accMarker = dl.genMarker[:common.HashLength]
|
accMarker = dl.genMarker[:common.HashLength]
|
||||||
}
|
}
|
||||||
|
|
@ -672,15 +677,11 @@ func (dl *diskLayer) generate(stats *generatorStats) {
|
||||||
defer ctx.close()
|
defer ctx.close()
|
||||||
|
|
||||||
if err := generateAccounts(ctx, dl, accMarker); err != nil {
|
if err := generateAccounts(ctx, dl, accMarker); err != nil {
|
||||||
// Extract the received interruption signal if exists
|
// Check if error was due to abort
|
||||||
if aerr, ok := err.(*abortErr); ok {
|
if err == errAborted {
|
||||||
abort = aerr.abort
|
stats.Log("Aborting state snapshot generation", dl.root, dl.genMarker)
|
||||||
}
|
}
|
||||||
// Aborted by internal error, wait the signal
|
dl.genStats = stats
|
||||||
if abort == nil {
|
|
||||||
abort = <-dl.genAbort
|
|
||||||
}
|
|
||||||
abort <- stats
|
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
// Snapshot fully generated, set the marker to nil.
|
// Snapshot fully generated, set the marker to nil.
|
||||||
|
|
@ -689,9 +690,7 @@ func (dl *diskLayer) generate(stats *generatorStats) {
|
||||||
journalProgress(ctx.batch, nil, stats)
|
journalProgress(ctx.batch, nil, stats)
|
||||||
if err := ctx.batch.Write(); err != nil {
|
if err := ctx.batch.Write(); err != nil {
|
||||||
log.Error("Failed to flush batch", "err", err)
|
log.Error("Failed to flush batch", "err", err)
|
||||||
|
dl.genStats = stats
|
||||||
abort = <-dl.genAbort
|
|
||||||
abort <- stats
|
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
ctx.batch.Reset()
|
ctx.batch.Reset()
|
||||||
|
|
@ -701,12 +700,9 @@ func (dl *diskLayer) generate(stats *generatorStats) {
|
||||||
|
|
||||||
dl.lock.Lock()
|
dl.lock.Lock()
|
||||||
dl.genMarker = nil
|
dl.genMarker = nil
|
||||||
|
dl.genStats = stats
|
||||||
close(dl.genPending)
|
close(dl.genPending)
|
||||||
dl.lock.Unlock()
|
dl.lock.Unlock()
|
||||||
|
|
||||||
// Someone will be looking for us, wait it out
|
|
||||||
abort = <-dl.genAbort
|
|
||||||
abort <- nil
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// increaseKey increase the input key by one bit. Return nil if the entire
|
// increaseKey increase the input key by one bit. Return nil if the entire
|
||||||
|
|
@ -720,17 +716,3 @@ func increaseKey(key []byte) []byte {
|
||||||
}
|
}
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// abortErr wraps an interruption signal received to represent the
|
|
||||||
// generation is aborted by external processes.
|
|
||||||
type abortErr struct {
|
|
||||||
abort chan *generatorStats
|
|
||||||
}
|
|
||||||
|
|
||||||
func newAbortErr(abort chan *generatorStats) error {
|
|
||||||
return &abortErr{abort: abort}
|
|
||||||
}
|
|
||||||
|
|
||||||
func (err *abortErr) Error() string {
|
|
||||||
return "aborted"
|
|
||||||
}
|
|
||||||
|
|
|
||||||
|
|
@ -79,10 +79,10 @@ func testGeneration(t *testing.T, scheme string) {
|
||||||
}
|
}
|
||||||
checkSnapRoot(t, snap, root)
|
checkSnapRoot(t, snap, root)
|
||||||
|
|
||||||
// Signal abortion to the generator and wait for it to tear down
|
// Stop the generator (if still running) and wait for it to exit.
|
||||||
stop := make(chan *generatorStats)
|
if err := snap.Release(); err != nil {
|
||||||
snap.genAbort <- stop
|
t.Fatal(err)
|
||||||
<-stop
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Tests that snapshot generation with existent flat state.
|
// Tests that snapshot generation with existent flat state.
|
||||||
|
|
@ -120,10 +120,10 @@ func testGenerateExistentState(t *testing.T, scheme string) {
|
||||||
}
|
}
|
||||||
checkSnapRoot(t, snap, root)
|
checkSnapRoot(t, snap, root)
|
||||||
|
|
||||||
// Signal abortion to the generator and wait for it to tear down
|
// Stop the generator (if still running) and wait for it to exit.
|
||||||
stop := make(chan *generatorStats)
|
if err := snap.Release(); err != nil {
|
||||||
snap.genAbort <- stop
|
t.Fatal(err)
|
||||||
<-stop
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func checkSnapRoot(t *testing.T, snap *diskLayer, trieRoot common.Hash) {
|
func checkSnapRoot(t *testing.T, snap *diskLayer, trieRoot common.Hash) {
|
||||||
|
|
@ -356,10 +356,10 @@ func testGenerateExistentStateWithWrongStorage(t *testing.T, scheme string) {
|
||||||
t.Errorf("Snapshot generation failed")
|
t.Errorf("Snapshot generation failed")
|
||||||
}
|
}
|
||||||
checkSnapRoot(t, snap, root)
|
checkSnapRoot(t, snap, root)
|
||||||
// Signal abortion to the generator and wait for it to tear down
|
// Stop the generator (if still running) and wait for it to exit.
|
||||||
stop := make(chan *generatorStats)
|
if err := snap.Release(); err != nil {
|
||||||
snap.genAbort <- stop
|
t.Fatal(err)
|
||||||
<-stop
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Tests that snapshot generation with existent flat state, where the flat state
|
// Tests that snapshot generation with existent flat state, where the flat state
|
||||||
|
|
@ -419,10 +419,10 @@ func testGenerateExistentStateWithWrongAccounts(t *testing.T, scheme string) {
|
||||||
}
|
}
|
||||||
checkSnapRoot(t, snap, root)
|
checkSnapRoot(t, snap, root)
|
||||||
|
|
||||||
// Signal abortion to the generator and wait for it to tear down
|
// Stop the generator (if still running) and wait for it to exit.
|
||||||
stop := make(chan *generatorStats)
|
if err := snap.Release(); err != nil {
|
||||||
snap.genAbort <- stop
|
t.Fatal(err)
|
||||||
<-stop
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Tests that snapshot generation errors out correctly in case of a missing trie
|
// Tests that snapshot generation errors out correctly in case of a missing trie
|
||||||
|
|
@ -459,10 +459,10 @@ func testGenerateCorruptAccountTrie(t *testing.T, scheme string) {
|
||||||
case <-time.After(time.Second):
|
case <-time.After(time.Second):
|
||||||
// Not generated fast enough, hopefully blocked inside on missing trie node fail
|
// Not generated fast enough, hopefully blocked inside on missing trie node fail
|
||||||
}
|
}
|
||||||
// Signal abortion to the generator and wait for it to tear down
|
// Stop the generator (if still running) and wait for it to exit.
|
||||||
stop := make(chan *generatorStats)
|
if err := snap.Release(); err != nil {
|
||||||
snap.genAbort <- stop
|
t.Fatal(err)
|
||||||
<-stop
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Tests that snapshot generation errors out correctly in case of a missing root
|
// Tests that snapshot generation errors out correctly in case of a missing root
|
||||||
|
|
@ -503,10 +503,10 @@ func testGenerateMissingStorageTrie(t *testing.T, scheme string) {
|
||||||
case <-time.After(time.Second):
|
case <-time.After(time.Second):
|
||||||
// Not generated fast enough, hopefully blocked inside on missing trie node fail
|
// Not generated fast enough, hopefully blocked inside on missing trie node fail
|
||||||
}
|
}
|
||||||
// Signal abortion to the generator and wait for it to tear down
|
// Stop the generator (if still running) and wait for it to exit.
|
||||||
stop := make(chan *generatorStats)
|
if err := snap.Release(); err != nil {
|
||||||
snap.genAbort <- stop
|
t.Fatal(err)
|
||||||
<-stop
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Tests that snapshot generation errors out correctly in case of a missing trie
|
// Tests that snapshot generation errors out correctly in case of a missing trie
|
||||||
|
|
@ -545,10 +545,10 @@ func testGenerateCorruptStorageTrie(t *testing.T, scheme string) {
|
||||||
case <-time.After(time.Second):
|
case <-time.After(time.Second):
|
||||||
// Not generated fast enough, hopefully blocked inside on missing trie node fail
|
// Not generated fast enough, hopefully blocked inside on missing trie node fail
|
||||||
}
|
}
|
||||||
// Signal abortion to the generator and wait for it to tear down
|
// Stop the generator (if still running) and wait for it to exit.
|
||||||
stop := make(chan *generatorStats)
|
if err := snap.Release(); err != nil {
|
||||||
snap.genAbort <- stop
|
t.Fatal(err)
|
||||||
<-stop
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Tests that snapshot generation when an extra account with storage exists in the snap state.
|
// Tests that snapshot generation when an extra account with storage exists in the snap state.
|
||||||
|
|
@ -610,10 +610,10 @@ func testGenerateWithExtraAccounts(t *testing.T, scheme string) {
|
||||||
}
|
}
|
||||||
checkSnapRoot(t, snap, root)
|
checkSnapRoot(t, snap, root)
|
||||||
|
|
||||||
// Signal abortion to the generator and wait for it to tear down
|
// Stop the generator (if still running) and wait for it to exit.
|
||||||
stop := make(chan *generatorStats)
|
if err := snap.Release(); err != nil {
|
||||||
snap.genAbort <- stop
|
t.Fatal(err)
|
||||||
<-stop
|
}
|
||||||
// If we now inspect the snap db, there should exist no extraneous storage items
|
// If we now inspect the snap db, there should exist no extraneous storage items
|
||||||
if data := rawdb.ReadStorageSnapshot(helper.diskdb, hashData([]byte("acc-2")), hashData([]byte("b-key-1"))); data != nil {
|
if data := rawdb.ReadStorageSnapshot(helper.diskdb, hashData([]byte("acc-2")), hashData([]byte("b-key-1"))); data != nil {
|
||||||
t.Fatalf("expected slot to be removed, got %v", string(data))
|
t.Fatalf("expected slot to be removed, got %v", string(data))
|
||||||
|
|
@ -671,10 +671,10 @@ func testGenerateWithManyExtraAccounts(t *testing.T, scheme string) {
|
||||||
t.Errorf("Snapshot generation failed")
|
t.Errorf("Snapshot generation failed")
|
||||||
}
|
}
|
||||||
checkSnapRoot(t, snap, root)
|
checkSnapRoot(t, snap, root)
|
||||||
// Signal abortion to the generator and wait for it to tear down
|
// Stop the generator (if still running) and wait for it to exit.
|
||||||
stop := make(chan *generatorStats)
|
if err := snap.Release(); err != nil {
|
||||||
snap.genAbort <- stop
|
t.Fatal(err)
|
||||||
<-stop
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Tests this case
|
// Tests this case
|
||||||
|
|
@ -720,10 +720,10 @@ func testGenerateWithExtraBeforeAndAfter(t *testing.T, scheme string) {
|
||||||
t.Errorf("Snapshot generation failed")
|
t.Errorf("Snapshot generation failed")
|
||||||
}
|
}
|
||||||
checkSnapRoot(t, snap, root)
|
checkSnapRoot(t, snap, root)
|
||||||
// Signal abortion to the generator and wait for it to tear down
|
// Stop the generator (if still running) and wait for it to exit.
|
||||||
stop := make(chan *generatorStats)
|
if err := snap.Release(); err != nil {
|
||||||
snap.genAbort <- stop
|
t.Fatal(err)
|
||||||
<-stop
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// TestGenerateWithMalformedSnapdata tests what happes if we have some junk
|
// TestGenerateWithMalformedSnapdata tests what happes if we have some junk
|
||||||
|
|
@ -760,10 +760,10 @@ func testGenerateWithMalformedSnapdata(t *testing.T, scheme string) {
|
||||||
t.Errorf("Snapshot generation failed")
|
t.Errorf("Snapshot generation failed")
|
||||||
}
|
}
|
||||||
checkSnapRoot(t, snap, root)
|
checkSnapRoot(t, snap, root)
|
||||||
// Signal abortion to the generator and wait for it to tear down
|
// Stop the generator (if still running) and wait for it to exit.
|
||||||
stop := make(chan *generatorStats)
|
if err := snap.Release(); err != nil {
|
||||||
snap.genAbort <- stop
|
t.Fatal(err)
|
||||||
<-stop
|
}
|
||||||
// If we now inspect the snap db, there should exist no extraneous storage items
|
// If we now inspect the snap db, there should exist no extraneous storage items
|
||||||
if data := rawdb.ReadStorageSnapshot(helper.diskdb, hashData([]byte("acc-2")), hashData([]byte("b-key-1"))); data != nil {
|
if data := rawdb.ReadStorageSnapshot(helper.diskdb, hashData([]byte("acc-2")), hashData([]byte("b-key-1"))); data != nil {
|
||||||
t.Fatalf("expected slot to be removed, got %v", string(data))
|
t.Fatalf("expected slot to be removed, got %v", string(data))
|
||||||
|
|
@ -797,10 +797,10 @@ func testGenerateFromEmptySnap(t *testing.T, scheme string) {
|
||||||
t.Errorf("Snapshot generation failed")
|
t.Errorf("Snapshot generation failed")
|
||||||
}
|
}
|
||||||
checkSnapRoot(t, snap, root)
|
checkSnapRoot(t, snap, root)
|
||||||
// Signal abortion to the generator and wait for it to tear down
|
// Stop the generator (if still running) and wait for it to exit.
|
||||||
stop := make(chan *generatorStats)
|
if err := snap.Release(); err != nil {
|
||||||
snap.genAbort <- stop
|
t.Fatal(err)
|
||||||
<-stop
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Tests that snapshot generation with existent flat state, where the flat state
|
// Tests that snapshot generation with existent flat state, where the flat state
|
||||||
|
|
@ -848,10 +848,10 @@ func testGenerateWithIncompleteStorage(t *testing.T, scheme string) {
|
||||||
t.Errorf("Snapshot generation failed")
|
t.Errorf("Snapshot generation failed")
|
||||||
}
|
}
|
||||||
checkSnapRoot(t, snap, root)
|
checkSnapRoot(t, snap, root)
|
||||||
// Signal abortion to the generator and wait for it to tear down
|
// Stop the generator (if still running) and wait for it to exit.
|
||||||
stop := make(chan *generatorStats)
|
if err := snap.Release(); err != nil {
|
||||||
snap.genAbort <- stop
|
t.Fatal(err)
|
||||||
<-stop
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func incKey(key []byte) []byte {
|
func incKey(key []byte) []byte {
|
||||||
|
|
@ -944,10 +944,10 @@ func testGenerateCompleteSnapshotWithDanglingStorage(t *testing.T, scheme string
|
||||||
}
|
}
|
||||||
checkSnapRoot(t, snap, root)
|
checkSnapRoot(t, snap, root)
|
||||||
|
|
||||||
// Signal abortion to the generator and wait for it to tear down
|
// Stop the generator (if still running) and wait for it to exit.
|
||||||
stop := make(chan *generatorStats)
|
if err := snap.Release(); err != nil {
|
||||||
snap.genAbort <- stop
|
t.Fatal(err)
|
||||||
<-stop
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Tests that snapshot generation with dangling storages. Dangling storage means
|
// Tests that snapshot generation with dangling storages. Dangling storage means
|
||||||
|
|
@ -981,10 +981,10 @@ func testGenerateBrokenSnapshotWithDanglingStorage(t *testing.T, scheme string)
|
||||||
}
|
}
|
||||||
checkSnapRoot(t, snap, root)
|
checkSnapRoot(t, snap, root)
|
||||||
|
|
||||||
// Signal abortion to the generator and wait for it to tear down
|
// Stop the generator (if still running) and wait for it to exit.
|
||||||
stop := make(chan *generatorStats)
|
if err := snap.Release(); err != nil {
|
||||||
snap.genAbort <- stop
|
t.Fatal(err)
|
||||||
<-stop
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// TestReleaseStopsGeneration verifies that Release() properly stops ongoing
|
// TestReleaseStopsGeneration verifies that Release() properly stops ongoing
|
||||||
|
|
|
||||||
|
|
@ -179,7 +179,8 @@ func loadSnapshot(diskdb ethdb.KeyValueStore, triedb *triedb.Database, root comm
|
||||||
// if the background generation is allowed
|
// if the background generation is allowed
|
||||||
if !generator.Done && !noBuild {
|
if !generator.Done && !noBuild {
|
||||||
base.genPending = make(chan struct{})
|
base.genPending = make(chan struct{})
|
||||||
base.genAbort = make(chan chan *generatorStats)
|
base.cancel = make(chan struct{})
|
||||||
|
base.done = make(chan struct{})
|
||||||
|
|
||||||
var origin uint64
|
var origin uint64
|
||||||
if len(generator.Marker) >= 8 {
|
if len(generator.Marker) >= 8 {
|
||||||
|
|
@ -199,16 +200,9 @@ func loadSnapshot(diskdb ethdb.KeyValueStore, triedb *triedb.Database, root comm
|
||||||
// Journal terminates any in-progress snapshot generation, also implicitly pushing
|
// Journal terminates any in-progress snapshot generation, also implicitly pushing
|
||||||
// the progress into the database.
|
// the progress into the database.
|
||||||
func (dl *diskLayer) Journal(buffer *bytes.Buffer) (common.Hash, error) {
|
func (dl *diskLayer) Journal(buffer *bytes.Buffer) (common.Hash, error) {
|
||||||
// If the snapshot is currently being generated, abort it
|
// If the snapshot is currently being generated, stop it
|
||||||
var stats *generatorStats
|
dl.stopGeneration()
|
||||||
if dl.genAbort != nil {
|
|
||||||
abort := make(chan *generatorStats)
|
|
||||||
dl.genAbort <- abort
|
|
||||||
|
|
||||||
if stats = <-abort; stats != nil {
|
|
||||||
stats.Log("Journalling in-progress snapshot", dl.root, dl.genMarker)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
// Ensure the layer didn't get stale
|
// Ensure the layer didn't get stale
|
||||||
dl.lock.RLock()
|
dl.lock.RLock()
|
||||||
defer dl.lock.RUnlock()
|
defer dl.lock.RUnlock()
|
||||||
|
|
@ -216,8 +210,8 @@ func (dl *diskLayer) Journal(buffer *bytes.Buffer) (common.Hash, error) {
|
||||||
if dl.stale {
|
if dl.stale {
|
||||||
return common.Hash{}, ErrSnapshotStale
|
return common.Hash{}, ErrSnapshotStale
|
||||||
}
|
}
|
||||||
// Ensure the generator stats is written even if none was ran this cycle
|
// Ensure the generator marker is written even if none was ran this cycle
|
||||||
journalProgress(dl.diskdb, dl.genMarker, stats)
|
journalProgress(dl.diskdb, dl.genMarker, nil)
|
||||||
|
|
||||||
log.Debug("Journalled disk layer", "root", dl.root)
|
log.Debug("Journalled disk layer", "root", dl.root)
|
||||||
return dl.root, nil
|
return dl.root, nil
|
||||||
|
|
|
||||||
|
|
@ -492,7 +492,7 @@ func (t *Tree) cap(diff *diffLayer, layers int) *diskLayer {
|
||||||
// there's a snapshot being generated currently. In that case, the trie
|
// there's a snapshot being generated currently. In that case, the trie
|
||||||
// will move from underneath the generator so we **must** merge all the
|
// will move from underneath the generator so we **must** merge all the
|
||||||
// partial data down into the snapshot and restart the generation.
|
// partial data down into the snapshot and restart the generation.
|
||||||
if flattened.parent.(*diskLayer).genAbort == nil {
|
if flattened.parent.(*diskLayer).cancel == nil {
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
@ -520,14 +520,10 @@ func diffToDisk(bottom *diffLayer) *diskLayer {
|
||||||
var (
|
var (
|
||||||
base = bottom.parent.(*diskLayer)
|
base = bottom.parent.(*diskLayer)
|
||||||
batch = base.diskdb.NewBatch()
|
batch = base.diskdb.NewBatch()
|
||||||
stats *generatorStats
|
|
||||||
)
|
)
|
||||||
// If the disk layer is running a snapshot generator, abort it
|
// Attempt to stop generation (if not already stopped)
|
||||||
if base.genAbort != nil {
|
base.stopGeneration()
|
||||||
abort := make(chan *generatorStats)
|
|
||||||
base.genAbort <- abort
|
|
||||||
stats = <-abort
|
|
||||||
}
|
|
||||||
// Put the deletion in the batch writer, flush all updates in the final step.
|
// Put the deletion in the batch writer, flush all updates in the final step.
|
||||||
rawdb.DeleteSnapshotRoot(batch)
|
rawdb.DeleteSnapshotRoot(batch)
|
||||||
|
|
||||||
|
|
@ -606,8 +602,8 @@ func diffToDisk(bottom *diffLayer) *diskLayer {
|
||||||
// Update the snapshot block marker and write any remainder data
|
// Update the snapshot block marker and write any remainder data
|
||||||
rawdb.WriteSnapshotRoot(batch, bottom.root)
|
rawdb.WriteSnapshotRoot(batch, bottom.root)
|
||||||
|
|
||||||
// Write out the generator progress marker and report
|
// Write out the generator progress marker
|
||||||
journalProgress(batch, base.genMarker, stats)
|
journalProgress(batch, base.genMarker, base.genStats)
|
||||||
|
|
||||||
// Flush all the updates in the single db operation. Ensure the
|
// Flush all the updates in the single db operation. Ensure the
|
||||||
// disk layer transition is atomic.
|
// disk layer transition is atomic.
|
||||||
|
|
@ -626,12 +622,13 @@ func diffToDisk(bottom *diffLayer) *diskLayer {
|
||||||
// If snapshot generation hasn't finished yet, port over all the starts and
|
// If snapshot generation hasn't finished yet, port over all the starts and
|
||||||
// continue where the previous round left off.
|
// continue where the previous round left off.
|
||||||
//
|
//
|
||||||
// Note, the `base.genAbort` comparison is not used normally, it's checked
|
// Note, the `base.genPending` comparison is not used normally, it's checked
|
||||||
// to allow the tests to play with the marker without triggering this path.
|
// to allow the tests to play with the marker without triggering this path.
|
||||||
if base.genMarker != nil && base.genAbort != nil {
|
if base.genMarker != nil && base.genPending != nil {
|
||||||
res.genMarker = base.genMarker
|
res.genMarker = base.genMarker
|
||||||
res.genAbort = make(chan chan *generatorStats)
|
res.cancel = make(chan struct{})
|
||||||
go res.generate(stats)
|
res.done = make(chan struct{})
|
||||||
|
go res.generate(base.genStats)
|
||||||
}
|
}
|
||||||
return res
|
return res
|
||||||
}
|
}
|
||||||
|
|
|
||||||
Loading…
Reference in a new issue