mirror of
https://github.com/ethereum/go-ethereum.git
synced 2026-04-24 18:42:23 +00:00
It's a PR based on #33303 and introduces an approach for trienode history indexing. --- In the current archive node design, resolving a historical trie node at a specific block involves the following steps: - Look up the corresponding trie node index and locate the first entry whose state ID is greater than the target state ID. - Resolve the trie node from the associated trienode history object. A naive approach would be to store mutation records for every trie node, similar to how flat state mutations are recorded. However, the total number of trie nodes is extremely large (approximately 2.4 billion), and the vast majority of them are rarely modified. Creating an index entry for each individual trie node would be very wasteful in both storage and indexing overhead. To address this, we aggregate multiple trie nodes into chunks and index mutations at the chunk level instead. --- For a storage trie, the trie is vertically partitioned into multiple sub tries, each spanning three consecutive levels. The top three levels (1 + 16 + 256 nodes) form the first chunk, and every subsequent three-level segment forms another chunk. ``` Original trie structure Level 0 [ ROOT ] 1 node Level 1 [0] [1] [2] ... [f] 16 nodes Level 2 [00] [01] ... [0f] [10] ... [ff] 256 nodes Level 3 [000] [001] ... [00f] [010] ... [fff] 4096 nodes Level 4 [0000] ... [000f] [0010] ... [001f] ... [ffff] 65536 nodes Vertical split into chunks (3 levels per chunk) Level0 [ ROOT ] 1 chunk Level3 [000] ... [fff] 4096 chunks Level6 [000000] ... [fffffff] 16777216 chunks ``` Within each chunk, there are 273 nodes in total, regardless of the chunk's depth in the trie. ``` Level 0 [ 0 ] 1 node Level 1 [ 1 ] … [ 16 ] 16 nodes Level 2 [ 17 ] … … [ 272 ] 256 nodes ``` Each chunk is uniquely identified by the path prefix of the root node of its corresponding sub-trie. Within a chunk, nodes are identified by a numeric index ranging from 0 to 272. For example, suppose that at block 100, the nodes with paths `[]`, `[0]`, `[f]`, `[00]`, and `[ff]` are modified. The mutation record for chunk 0 is then appended with the following entry: `[100 → [0, 1, 16, 17, 272]]`, `272` is the numeric ID of path `[ff]`. Furthermore, due to the structural properties of the Merkle Patricia Trie, if a child node is modified, all of its ancestors along the same path must also be updated. As a result, in the above example, recording mutations for nodes `00` and `ff` alone is sufficient, as this implicitly indicates that their ancestor nodes `[]`, `[0]` and `[f]` were also modified at block 100. --- Query processing is slightly more complicated. Since trie nodes are indexed at the chunk level, each individual trie node lookup requires an additional filtering step to ensure that a given mutation record actually corresponds to the target trie node. As mentioned earlier, mutation records store only the numeric identifiers of leaf nodes, while ancestor nodes are omitted for storage efficiency. Consequently, when querying an ancestor node, additional checks are required to determine whether the mutation record implicitly represents a modification to that ancestor. Moreover, since trie nodes are indexed at the chunk level, some trie nodes may be updated frequently, causing their mutation records to dominate the index. Queries targeting rarely modified trie nodes would then scan a large amount of irrelevant index data, significantly degrading performance. To address this issue, a bitmap is introduced for each index block and stored in the chunk's metadata. Before loading a specific index block, the bitmap is checked to determine whether the block contains mutation records relevant to the target trie node. If the bitmap indicates that the block does not contain such records, the block is skipped entirely.
487 lines
16 KiB
Go
487 lines
16 KiB
Go
// Copyright 2025 The go-ethereum Authors
|
|
// This file is part of the go-ethereum library.
|
|
//
|
|
// The go-ethereum library is free software: you can redistribute it and/or modify
|
|
// it under the terms of the GNU Lesser General Public License as published by
|
|
// the Free Software Foundation, either version 3 of the License, or
|
|
// (at your option) any later version.
|
|
//
|
|
// The go-ethereum library is distributed in the hope that it will be useful,
|
|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
// GNU Lesser General Public License for more details.
|
|
//
|
|
// You should have received a copy of the GNU Lesser General Public License
|
|
// along with the go-ethereum library. If not, see <http://www.gnu.org/licenses/
|
|
|
|
package pathdb
|
|
|
|
import (
|
|
"errors"
|
|
"fmt"
|
|
"math"
|
|
|
|
"github.com/ethereum/go-ethereum/core/rawdb"
|
|
"github.com/ethereum/go-ethereum/ethdb"
|
|
)
|
|
|
|
// parseIndex parses the index data from the provided byte stream. The index data
|
|
// is a sequence of fixed-size metadata entries, and any empty metadata entry is
|
|
// considered invalid.
|
|
//
|
|
// Each metadata entry consists of two components: the indexBlockDesc and an
|
|
// optional extension bitmap. The bitmap length may vary across different categories,
|
|
// but must remain consistent within the same category.
|
|
func parseIndex(blob []byte, bitmapSize int) ([]*indexBlockDesc, error) {
|
|
if len(blob) == 0 {
|
|
return nil, errors.New("empty state history index")
|
|
}
|
|
size := indexBlockDescSize + bitmapSize
|
|
if len(blob)%size != 0 {
|
|
return nil, fmt.Errorf("corrupted state index, len: %d, bitmap size: %d", len(blob), bitmapSize)
|
|
}
|
|
var (
|
|
lastID uint32
|
|
descList []*indexBlockDesc
|
|
)
|
|
for i := 0; i < len(blob)/size; i++ {
|
|
var desc indexBlockDesc
|
|
desc.decode(blob[i*size : (i+1)*size])
|
|
if desc.empty() {
|
|
return nil, errors.New("empty state history index block")
|
|
}
|
|
if lastID != 0 {
|
|
if lastID+1 != desc.id {
|
|
return nil, fmt.Errorf("index block id is out of order, last-id: %d, this-id: %d", lastID, desc.id)
|
|
}
|
|
// Theoretically, order should be validated between consecutive index blocks,
|
|
// ensuring that elements within them are strictly ordered. However, since
|
|
// tracking the minimum element in each block has non-trivial storage overhead,
|
|
// this check is optimistically omitted.
|
|
//
|
|
// TODO(rjl493456442) the minimal element can be resolved from the index block,
|
|
// evaluate the check cost (mostly IO overhead).
|
|
|
|
/* if desc.min <= lastMax {
|
|
return nil, fmt.Errorf("index block range is out of order, last-max: %d, this-min: %d", lastMax, desc.min)
|
|
}*/
|
|
}
|
|
lastID = desc.id
|
|
descList = append(descList, &desc)
|
|
}
|
|
return descList, nil
|
|
}
|
|
|
|
// indexReader is the structure to look up the state history index records
|
|
// associated with the specific state element.
|
|
type indexReader struct {
|
|
db ethdb.KeyValueReader
|
|
descList []*indexBlockDesc
|
|
readers map[uint32]*blockReader
|
|
state stateIdent
|
|
bitmapSize int
|
|
}
|
|
|
|
// loadIndexData loads the index data associated with the specified state.
|
|
func loadIndexData(db ethdb.KeyValueReader, state stateIdent, bitmapSize int) ([]*indexBlockDesc, error) {
|
|
blob := readStateIndex(state, db)
|
|
if len(blob) == 0 {
|
|
return nil, nil
|
|
}
|
|
return parseIndex(blob, bitmapSize)
|
|
}
|
|
|
|
// newIndexReader constructs a index reader for the specified state. Reader with
|
|
// empty data is allowed.
|
|
func newIndexReader(db ethdb.KeyValueReader, state stateIdent, bitmapSize int) (*indexReader, error) {
|
|
descList, err := loadIndexData(db, state, bitmapSize)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
return &indexReader{
|
|
descList: descList,
|
|
readers: make(map[uint32]*blockReader),
|
|
db: db,
|
|
state: state,
|
|
bitmapSize: bitmapSize,
|
|
}, nil
|
|
}
|
|
|
|
// refresh reloads the last section of index data to account for any additional
|
|
// elements that may have been written to disk.
|
|
func (r *indexReader) refresh() error {
|
|
// Release the reader for the last section of index data, as its content
|
|
// may have been modified by additional elements written to the disk.
|
|
if len(r.descList) != 0 {
|
|
last := r.descList[len(r.descList)-1]
|
|
delete(r.readers, last.id)
|
|
}
|
|
descList, err := loadIndexData(r.db, r.state, r.bitmapSize)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
r.descList = descList
|
|
return nil
|
|
}
|
|
|
|
// readGreaterThan locates the first element that is greater than the specified
|
|
// id. If no such element is found, MaxUint64 is returned.
|
|
func (r *indexReader) readGreaterThan(id uint64) (uint64, error) {
|
|
it := r.newIterator(nil)
|
|
found := it.SeekGT(id)
|
|
if err := it.Error(); err != nil {
|
|
return 0, err
|
|
}
|
|
if !found {
|
|
return math.MaxUint64, nil
|
|
}
|
|
return it.ID(), nil
|
|
}
|
|
|
|
// indexWriter is responsible for writing index data for a specific state (either
|
|
// an account or a storage slot). The state index follows a two-layer structure:
|
|
// the first layer consists of a list of fixed-size metadata, each linked to a
|
|
// second-layer block. The index data (monotonically increasing list of state
|
|
// history ids) is stored in these second-layer index blocks, which are size
|
|
// limited.
|
|
type indexWriter struct {
|
|
descList []*indexBlockDesc // The list of index block descriptions
|
|
bw *blockWriter // The live index block writer
|
|
frozen []*blockWriter // The finalized index block writers, waiting for flush
|
|
lastID uint64 // The ID of the latest tracked history
|
|
state stateIdent // The identifier of the state being indexed
|
|
bitmapSize int // The size of optional extension bitmap
|
|
db ethdb.KeyValueReader
|
|
}
|
|
|
|
// newIndexWriter constructs the index writer for the specified state. Additionally,
|
|
// it takes an integer as the limit and prunes all existing elements above that ID.
|
|
// It's essential as the recovery mechanism after unclean shutdown during the history
|
|
// indexing.
|
|
func newIndexWriter(db ethdb.KeyValueReader, state stateIdent, limit uint64, bitmapSize int) (*indexWriter, error) {
|
|
blob := readStateIndex(state, db)
|
|
if len(blob) == 0 {
|
|
desc := newIndexBlockDesc(0, bitmapSize)
|
|
bw, _ := newBlockWriter(nil, desc, 0 /* useless if the block is empty */, bitmapSize != 0)
|
|
return &indexWriter{
|
|
descList: []*indexBlockDesc{desc},
|
|
bw: bw,
|
|
state: state,
|
|
db: db,
|
|
bitmapSize: bitmapSize,
|
|
}, nil
|
|
}
|
|
descList, err := parseIndex(blob, bitmapSize)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
// Trim trailing blocks whose elements all exceed the limit.
|
|
for i := len(descList) - 1; i > 0 && descList[i].max > limit; i-- {
|
|
// The previous block has the elements that exceed the limit,
|
|
// therefore the current block can be entirely dropped.
|
|
if descList[i-1].max >= limit {
|
|
descList = descList[:i]
|
|
}
|
|
}
|
|
// Take the last block for appending new elements
|
|
lastDesc := descList[len(descList)-1]
|
|
indexBlock := readStateIndexBlock(state, db, lastDesc.id)
|
|
|
|
// Construct the writer for the last block. All elements in this block
|
|
// that exceed the limit will be truncated.
|
|
bw, err := newBlockWriter(indexBlock, lastDesc, limit, bitmapSize != 0)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
return &indexWriter{
|
|
descList: descList,
|
|
lastID: bw.last(),
|
|
bw: bw,
|
|
state: state,
|
|
db: db,
|
|
bitmapSize: bitmapSize,
|
|
}, nil
|
|
}
|
|
|
|
// append adds the new element into the index writer.
|
|
func (w *indexWriter) append(id uint64, ext []uint16) error {
|
|
if id <= w.lastID {
|
|
return fmt.Errorf("append element out of order, last: %d, this: %d", w.lastID, id)
|
|
}
|
|
if w.bw.estimateFull(ext) {
|
|
if err := w.rotate(); err != nil {
|
|
return err
|
|
}
|
|
}
|
|
if err := w.bw.append(id, ext); err != nil {
|
|
return err
|
|
}
|
|
w.lastID = id
|
|
|
|
return nil
|
|
}
|
|
|
|
// rotate creates a new index block for storing index records from scratch
|
|
// and caches the current full index block for finalization.
|
|
func (w *indexWriter) rotate() error {
|
|
var (
|
|
err error
|
|
desc = newIndexBlockDesc(w.bw.desc.id+1, w.bitmapSize)
|
|
)
|
|
w.frozen = append(w.frozen, w.bw)
|
|
w.bw, err = newBlockWriter(nil, desc, 0 /* useless if the block is empty */, w.bitmapSize != 0)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
w.descList = append(w.descList, desc)
|
|
return nil
|
|
}
|
|
|
|
// finish finalizes all the frozen index block writers along with the live one
|
|
// if it's not empty, committing the index block data and the index meta into
|
|
// the supplied batch.
|
|
//
|
|
// This function is safe to be called multiple times.
|
|
func (w *indexWriter) finish(batch ethdb.Batch) {
|
|
var (
|
|
writers = append(w.frozen, w.bw)
|
|
descList = w.descList
|
|
)
|
|
// The live index block writer might be empty if the entire index write
|
|
// is created from scratch, remove it from committing.
|
|
if w.bw.empty() {
|
|
writers = writers[:len(writers)-1]
|
|
descList = descList[:len(descList)-1]
|
|
}
|
|
if len(writers) == 0 {
|
|
return // nothing to commit
|
|
}
|
|
for _, bw := range writers {
|
|
writeStateIndexBlock(w.state, batch, bw.desc.id, bw.finish())
|
|
}
|
|
w.frozen = nil // release all the frozen writers
|
|
|
|
size := indexBlockDescSize + w.bitmapSize
|
|
buf := make([]byte, 0, size*len(descList))
|
|
for _, desc := range descList {
|
|
buf = append(buf, desc.encode()...)
|
|
}
|
|
writeStateIndex(w.state, batch, buf)
|
|
}
|
|
|
|
// indexDeleter is responsible for deleting index data for a specific state.
|
|
type indexDeleter struct {
|
|
descList []*indexBlockDesc // The list of index block descriptions
|
|
bw *blockWriter // The live index block writer
|
|
dropped []uint32 // The list of index block id waiting for deleting
|
|
lastID uint64 // The ID of the latest tracked history
|
|
state stateIdent // The identifier of the state being indexed
|
|
bitmapSize int // The size of optional extension bitmap
|
|
db ethdb.KeyValueReader
|
|
}
|
|
|
|
// newIndexDeleter constructs the index deleter for the specified state.
|
|
func newIndexDeleter(db ethdb.KeyValueReader, state stateIdent, limit uint64, bitmapSize int) (*indexDeleter, error) {
|
|
blob := readStateIndex(state, db)
|
|
if len(blob) == 0 {
|
|
// TODO(rjl493456442) we can probably return an error here,
|
|
// deleter with no data is meaningless.
|
|
desc := newIndexBlockDesc(0, bitmapSize)
|
|
bw, _ := newBlockWriter(nil, desc, 0 /* useless if the block is empty */, bitmapSize != 0)
|
|
return &indexDeleter{
|
|
descList: []*indexBlockDesc{desc},
|
|
bw: bw,
|
|
state: state,
|
|
bitmapSize: bitmapSize,
|
|
db: db,
|
|
}, nil
|
|
}
|
|
descList, err := parseIndex(blob, bitmapSize)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
// Trim trailing blocks whose elements all exceed the limit.
|
|
for i := len(descList) - 1; i > 0 && descList[i].max > limit; i-- {
|
|
// The previous block has the elements that exceed the limit,
|
|
// therefore the current block can be entirely dropped.
|
|
if descList[i-1].max >= limit {
|
|
descList = descList[:i]
|
|
}
|
|
}
|
|
// Take the block for deleting element from
|
|
lastDesc := descList[len(descList)-1]
|
|
indexBlock := readStateIndexBlock(state, db, lastDesc.id)
|
|
|
|
// Construct the writer for the last block. All elements in this block
|
|
// that exceed the limit will be truncated.
|
|
bw, err := newBlockWriter(indexBlock, lastDesc, limit, bitmapSize != 0)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
return &indexDeleter{
|
|
descList: descList,
|
|
lastID: bw.last(),
|
|
bw: bw,
|
|
state: state,
|
|
bitmapSize: bitmapSize,
|
|
db: db,
|
|
}, nil
|
|
}
|
|
|
|
// empty returns whether the state index is empty.
|
|
func (d *indexDeleter) empty() bool {
|
|
return d.bw.empty() && len(d.descList) == 1
|
|
}
|
|
|
|
// pop removes the last written element from the index writer.
|
|
func (d *indexDeleter) pop(id uint64) error {
|
|
if id == 0 {
|
|
return errors.New("zero history ID is not valid")
|
|
}
|
|
if id != d.lastID {
|
|
return fmt.Errorf("pop element out of order, last: %d, this: %d", d.lastID, id)
|
|
}
|
|
if err := d.bw.pop(id); err != nil {
|
|
return err
|
|
}
|
|
if !d.bw.empty() {
|
|
d.lastID = d.bw.desc.max
|
|
return nil
|
|
}
|
|
// Discarding the last block writer if it becomes empty by popping an element
|
|
d.dropped = append(d.dropped, d.descList[len(d.descList)-1].id)
|
|
|
|
// Reset the entire index writer if it becomes empty after popping an element
|
|
if d.empty() {
|
|
d.lastID = 0
|
|
return nil
|
|
}
|
|
d.descList = d.descList[:len(d.descList)-1]
|
|
|
|
// Open the previous block writer for deleting
|
|
lastDesc := d.descList[len(d.descList)-1]
|
|
indexBlock := readStateIndexBlock(d.state, d.db, lastDesc.id)
|
|
bw, err := newBlockWriter(indexBlock, lastDesc, lastDesc.max, d.bitmapSize != 0)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
d.bw = bw
|
|
d.lastID = bw.desc.max
|
|
return nil
|
|
}
|
|
|
|
// finish deletes the empty index blocks and updates the index meta.
|
|
//
|
|
// This function is safe to be called multiple times.
|
|
func (d *indexDeleter) finish(batch ethdb.Batch) {
|
|
for _, id := range d.dropped {
|
|
deleteStateIndexBlock(d.state, batch, id)
|
|
}
|
|
d.dropped = nil
|
|
|
|
// Flush the content of last block writer, regardless it's dirty or not
|
|
if !d.bw.empty() {
|
|
writeStateIndexBlock(d.state, batch, d.bw.desc.id, d.bw.finish())
|
|
}
|
|
// Flush the index metadata into the supplied batch
|
|
if d.empty() {
|
|
deleteStateIndex(d.state, batch)
|
|
} else {
|
|
size := indexBlockDescSize + d.bitmapSize
|
|
buf := make([]byte, 0, size*len(d.descList))
|
|
for _, desc := range d.descList {
|
|
buf = append(buf, desc.encode()...)
|
|
}
|
|
writeStateIndex(d.state, batch, buf)
|
|
}
|
|
}
|
|
|
|
// readStateIndex retrieves the index metadata for the given state identifier.
|
|
// This function is shared by accounts and storage slots.
|
|
func readStateIndex(ident stateIdent, db ethdb.KeyValueReader) []byte {
|
|
switch ident.typ {
|
|
case typeAccount:
|
|
return rawdb.ReadAccountHistoryIndex(db, ident.addressHash)
|
|
case typeStorage:
|
|
return rawdb.ReadStorageHistoryIndex(db, ident.addressHash, ident.storageHash)
|
|
case typeTrienode:
|
|
return rawdb.ReadTrienodeHistoryIndex(db, ident.addressHash, []byte(ident.path))
|
|
default:
|
|
panic(fmt.Errorf("unknown type: %v", ident.typ))
|
|
}
|
|
}
|
|
|
|
// writeStateIndex writes the provided index metadata into database with the
|
|
// given state identifier. This function is shared by accounts and storage slots.
|
|
func writeStateIndex(ident stateIdent, db ethdb.KeyValueWriter, data []byte) {
|
|
switch ident.typ {
|
|
case typeAccount:
|
|
rawdb.WriteAccountHistoryIndex(db, ident.addressHash, data)
|
|
case typeStorage:
|
|
rawdb.WriteStorageHistoryIndex(db, ident.addressHash, ident.storageHash, data)
|
|
case typeTrienode:
|
|
rawdb.WriteTrienodeHistoryIndex(db, ident.addressHash, []byte(ident.path), data)
|
|
default:
|
|
panic(fmt.Errorf("unknown type: %v", ident.typ))
|
|
}
|
|
}
|
|
|
|
// deleteStateIndex removes the index metadata for the given state identifier.
|
|
// This function is shared by accounts and storage slots.
|
|
func deleteStateIndex(ident stateIdent, db ethdb.KeyValueWriter) {
|
|
switch ident.typ {
|
|
case typeAccount:
|
|
rawdb.DeleteAccountHistoryIndex(db, ident.addressHash)
|
|
case typeStorage:
|
|
rawdb.DeleteStorageHistoryIndex(db, ident.addressHash, ident.storageHash)
|
|
case typeTrienode:
|
|
rawdb.DeleteTrienodeHistoryIndex(db, ident.addressHash, []byte(ident.path))
|
|
default:
|
|
panic(fmt.Errorf("unknown type: %v", ident.typ))
|
|
}
|
|
}
|
|
|
|
// readStateIndexBlock retrieves the index block for the given state identifier
|
|
// and block ID. This function is shared by accounts and storage slots.
|
|
func readStateIndexBlock(ident stateIdent, db ethdb.KeyValueReader, id uint32) []byte {
|
|
switch ident.typ {
|
|
case typeAccount:
|
|
return rawdb.ReadAccountHistoryIndexBlock(db, ident.addressHash, id)
|
|
case typeStorage:
|
|
return rawdb.ReadStorageHistoryIndexBlock(db, ident.addressHash, ident.storageHash, id)
|
|
case typeTrienode:
|
|
return rawdb.ReadTrienodeHistoryIndexBlock(db, ident.addressHash, []byte(ident.path), id)
|
|
default:
|
|
panic(fmt.Errorf("unknown type: %v", ident.typ))
|
|
}
|
|
}
|
|
|
|
// writeStateIndexBlock writes the provided index block into database with the
|
|
// given state identifier. This function is shared by accounts and storage slots.
|
|
func writeStateIndexBlock(ident stateIdent, db ethdb.KeyValueWriter, id uint32, data []byte) {
|
|
switch ident.typ {
|
|
case typeAccount:
|
|
rawdb.WriteAccountHistoryIndexBlock(db, ident.addressHash, id, data)
|
|
case typeStorage:
|
|
rawdb.WriteStorageHistoryIndexBlock(db, ident.addressHash, ident.storageHash, id, data)
|
|
case typeTrienode:
|
|
rawdb.WriteTrienodeHistoryIndexBlock(db, ident.addressHash, []byte(ident.path), id, data)
|
|
default:
|
|
panic(fmt.Errorf("unknown type: %v", ident.typ))
|
|
}
|
|
}
|
|
|
|
// deleteStateIndexBlock removes the index block from database with the given
|
|
// state identifier. This function is shared by accounts and storage slots.
|
|
func deleteStateIndexBlock(ident stateIdent, db ethdb.KeyValueWriter, id uint32) {
|
|
switch ident.typ {
|
|
case typeAccount:
|
|
rawdb.DeleteAccountHistoryIndexBlock(db, ident.addressHash, id)
|
|
case typeStorage:
|
|
rawdb.DeleteStorageHistoryIndexBlock(db, ident.addressHash, ident.storageHash, id)
|
|
case typeTrienode:
|
|
rawdb.DeleteTrienodeHistoryIndexBlock(db, ident.addressHash, []byte(ident.path), id)
|
|
default:
|
|
panic(fmt.Errorf("unknown type: %v", ident.typ))
|
|
}
|
|
}
|