mirror of
https://github.com/ethereum/go-ethereum.git
synced 2026-03-04 18:35:03 +00:00
Compare commits
52 commits
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
fc8c10476d | ||
|
|
402c71f2e2 | ||
|
|
28dad943f6 | ||
|
|
6d0dd08860 | ||
|
|
dd202d4283 | ||
|
|
814edc5308 | ||
|
|
6d99759f01 | ||
|
|
fe3a74e610 | ||
|
|
4f75049ea0 | ||
|
|
773f71bb9e | ||
|
|
856e4d55d8 | ||
|
|
db7d3a4e0e | ||
|
|
16783c167c | ||
|
|
9962e2c9f3 | ||
|
|
d318e8eba9 | ||
|
|
b25080cac0 | ||
|
|
48cfc97776 | ||
|
|
1eead2ec33 | ||
|
|
2726c9ef9e | ||
|
|
5695fbc156 | ||
|
|
825436f043 | ||
|
|
723aae2b4e | ||
|
|
cee751a1ed | ||
|
|
95c6b05806 | ||
|
|
7793e00f0d | ||
|
|
1b1133d669 | ||
|
|
be92f5487e | ||
|
|
8a4345611d | ||
|
|
f811bfe4fd | ||
|
|
406a852ec8 | ||
|
|
2a45272408 | ||
|
|
8450e40798 | ||
|
|
9ecb6c4ae6 | ||
|
|
e636e4e3c1 | ||
|
|
cbf3d8fed2 | ||
|
|
199ac16e07 | ||
|
|
01083736c8 | ||
|
|
59ad40e562 | ||
|
|
c2e1785a48 | ||
|
|
82fad31540 | ||
|
|
1625064c68 | ||
|
|
1d1a094d51 | ||
|
|
e40aa46e88 | ||
|
|
d3dd48e59d | ||
|
|
00cbd2e6f4 | ||
|
|
453d0f9299 | ||
|
|
6d865ccd30 | ||
|
|
54f72c796f | ||
|
|
2a62df3815 | ||
|
|
01fe1d716c | ||
|
|
3eed0580d4 | ||
|
|
1054276906 |
168 changed files with 4106 additions and 1539 deletions
1
.github/CODEOWNERS
vendored
1
.github/CODEOWNERS
vendored
|
|
@ -10,6 +10,7 @@ beacon/merkle/ @zsfelfoldi
|
||||||
beacon/types/ @zsfelfoldi @fjl
|
beacon/types/ @zsfelfoldi @fjl
|
||||||
beacon/params/ @zsfelfoldi @fjl
|
beacon/params/ @zsfelfoldi @fjl
|
||||||
cmd/evm/ @MariusVanDerWijden @lightclient
|
cmd/evm/ @MariusVanDerWijden @lightclient
|
||||||
|
cmd/keeper/ @gballet
|
||||||
core/state/ @rjl493456442
|
core/state/ @rjl493456442
|
||||||
crypto/ @gballet @jwasinger @fjl
|
crypto/ @gballet @jwasinger @fjl
|
||||||
core/ @rjl493456442
|
core/ @rjl493456442
|
||||||
|
|
|
||||||
4
.github/workflows/go.yml
vendored
4
.github/workflows/go.yml
vendored
|
|
@ -69,8 +69,8 @@ jobs:
|
||||||
|
|
||||||
- name: Install cross toolchain
|
- name: Install cross toolchain
|
||||||
run: |
|
run: |
|
||||||
apt-get update
|
sudo apt-get update
|
||||||
apt-get -yq --no-install-suggests --no-install-recommends install gcc-multilib
|
sudo apt-get -yq --no-install-suggests --no-install-recommends install gcc-multilib
|
||||||
|
|
||||||
- name: Build
|
- name: Build
|
||||||
run: go run build/ci.go test -arch 386 -short -p 8
|
run: go run build/ci.go test -arch 386 -short -p 8
|
||||||
|
|
|
||||||
102
AGENTS.md
Normal file
102
AGENTS.md
Normal file
|
|
@ -0,0 +1,102 @@
|
||||||
|
# AGENTS
|
||||||
|
|
||||||
|
## Guidelines
|
||||||
|
|
||||||
|
- **Keep changes minimal and focused.** Only modify code directly related to the task at hand. Do not refactor unrelated code, rename existing variables or functions for style, or bundle unrelated fixes into the same commit or PR.
|
||||||
|
- **Do not add, remove, or update dependencies** unless the task explicitly requires it.
|
||||||
|
|
||||||
|
## Pre-Commit Checklist
|
||||||
|
|
||||||
|
Before every commit, run **all** of the following checks and ensure they pass:
|
||||||
|
|
||||||
|
### 1. Formatting
|
||||||
|
|
||||||
|
Before committing, always run `gofmt` and `goimports` on all modified files:
|
||||||
|
|
||||||
|
```sh
|
||||||
|
gofmt -w <modified files>
|
||||||
|
goimports -w <modified files>
|
||||||
|
```
|
||||||
|
|
||||||
|
### 2. Build All Commands
|
||||||
|
|
||||||
|
Verify that all tools compile successfully:
|
||||||
|
|
||||||
|
```sh
|
||||||
|
make all
|
||||||
|
```
|
||||||
|
|
||||||
|
This builds all executables under `cmd/`, including `keeper` which has special build requirements.
|
||||||
|
|
||||||
|
### 3. Tests
|
||||||
|
|
||||||
|
While iterating during development, use `-short` for faster feedback:
|
||||||
|
|
||||||
|
```sh
|
||||||
|
go run ./build/ci.go test -short
|
||||||
|
```
|
||||||
|
|
||||||
|
Before committing, run the full test suite **without** `-short` to ensure all tests pass, including the Ethereum execution-spec tests and all state/block test permutations:
|
||||||
|
|
||||||
|
```sh
|
||||||
|
go run ./build/ci.go test
|
||||||
|
```
|
||||||
|
|
||||||
|
### 4. Linting
|
||||||
|
|
||||||
|
```sh
|
||||||
|
go run ./build/ci.go lint
|
||||||
|
```
|
||||||
|
|
||||||
|
This runs additional style checks. Fix any issues before committing.
|
||||||
|
|
||||||
|
### 5. Generated Code
|
||||||
|
|
||||||
|
```sh
|
||||||
|
go run ./build/ci.go check_generate
|
||||||
|
```
|
||||||
|
|
||||||
|
Ensures that all generated files (e.g., `gen_*.go`) are up to date. If this fails, first install the required code generators by running `make devtools`, then run the appropriate `go generate` commands and include the updated files in your commit.
|
||||||
|
|
||||||
|
### 6. Dependency Hygiene
|
||||||
|
|
||||||
|
```sh
|
||||||
|
go run ./build/ci.go check_baddeps
|
||||||
|
```
|
||||||
|
|
||||||
|
Verifies that no forbidden dependencies have been introduced.
|
||||||
|
|
||||||
|
## What to include in commits
|
||||||
|
|
||||||
|
Do not commit binaries, whether they are produced by the main build or byproducts of investigations.
|
||||||
|
|
||||||
|
## Commit Message Format
|
||||||
|
|
||||||
|
Commit messages must be prefixed with the package(s) they modify, followed by a short lowercase description:
|
||||||
|
|
||||||
|
```
|
||||||
|
<package(s)>: description
|
||||||
|
```
|
||||||
|
|
||||||
|
Examples:
|
||||||
|
- `core/vm: fix stack overflow in PUSH instruction`
|
||||||
|
- `eth, rpc: make trace configs optional`
|
||||||
|
- `cmd/geth: add new flag for sync mode`
|
||||||
|
|
||||||
|
Use comma-separated package names when multiple areas are affected. Keep the description concise.
|
||||||
|
|
||||||
|
## Pull Request Title Format
|
||||||
|
|
||||||
|
PR titles follow the same convention as commit messages:
|
||||||
|
|
||||||
|
```
|
||||||
|
<list of modified paths>: description
|
||||||
|
```
|
||||||
|
|
||||||
|
Examples:
|
||||||
|
- `core/vm: fix stack overflow in PUSH instruction`
|
||||||
|
- `core, eth: add arena allocator support`
|
||||||
|
- `cmd/geth, internal/ethapi: refactor transaction args`
|
||||||
|
- `trie/archiver: streaming subtree archival to fix OOM`
|
||||||
|
|
||||||
|
Use the top-level package paths, comma-separated if multiple areas are affected. Only mention the directories with functional changes, interface changes that trickle all over the codebase should not generate an exhaustive list. The description should be a short, lowercase summary of the change.
|
||||||
|
|
@ -4,7 +4,7 @@ ARG VERSION=""
|
||||||
ARG BUILDNUM=""
|
ARG BUILDNUM=""
|
||||||
|
|
||||||
# Build Geth in a stock Go builder container
|
# Build Geth in a stock Go builder container
|
||||||
FROM golang:1.24-alpine AS builder
|
FROM golang:1.26-alpine AS builder
|
||||||
|
|
||||||
RUN apk add --no-cache gcc musl-dev linux-headers git
|
RUN apk add --no-cache gcc musl-dev linux-headers git
|
||||||
|
|
||||||
|
|
|
||||||
|
|
@ -4,7 +4,7 @@ ARG VERSION=""
|
||||||
ARG BUILDNUM=""
|
ARG BUILDNUM=""
|
||||||
|
|
||||||
# Build Geth in a stock Go builder container
|
# Build Geth in a stock Go builder container
|
||||||
FROM golang:1.24-alpine AS builder
|
FROM golang:1.26-alpine AS builder
|
||||||
|
|
||||||
RUN apk add --no-cache gcc musl-dev linux-headers git
|
RUN apk add --no-cache gcc musl-dev linux-headers git
|
||||||
|
|
||||||
|
|
|
||||||
|
|
@ -87,6 +87,10 @@ func (ec *engineClient) updateLoop(headCh <-chan types.ChainHeadEvent) {
|
||||||
if status, err := ec.callForkchoiceUpdated(forkName, event); err == nil {
|
if status, err := ec.callForkchoiceUpdated(forkName, event); err == nil {
|
||||||
log.Info("Successful ForkchoiceUpdated", "head", event.Block.Hash(), "status", status)
|
log.Info("Successful ForkchoiceUpdated", "head", event.Block.Hash(), "status", status)
|
||||||
} else {
|
} else {
|
||||||
|
if err.Error() == "beacon syncer reorging" {
|
||||||
|
log.Debug("Failed ForkchoiceUpdated", "head", event.Block.Hash(), "error", err)
|
||||||
|
continue // ignore beacon syncer reorging errors, this error can occur if the blsync is skipping a block
|
||||||
|
}
|
||||||
log.Error("Failed ForkchoiceUpdated", "head", event.Block.Hash(), "error", err)
|
log.Error("Failed ForkchoiceUpdated", "head", event.Block.Hash(), "error", err)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
|
||||||
|
|
@ -21,6 +21,7 @@ func (p PayloadAttributes) MarshalJSON() ([]byte, error) {
|
||||||
SuggestedFeeRecipient common.Address `json:"suggestedFeeRecipient" gencodec:"required"`
|
SuggestedFeeRecipient common.Address `json:"suggestedFeeRecipient" gencodec:"required"`
|
||||||
Withdrawals []*types.Withdrawal `json:"withdrawals"`
|
Withdrawals []*types.Withdrawal `json:"withdrawals"`
|
||||||
BeaconRoot *common.Hash `json:"parentBeaconBlockRoot"`
|
BeaconRoot *common.Hash `json:"parentBeaconBlockRoot"`
|
||||||
|
SlotNumber *hexutil.Uint64 `json:"slotNumber"`
|
||||||
}
|
}
|
||||||
var enc PayloadAttributes
|
var enc PayloadAttributes
|
||||||
enc.Timestamp = hexutil.Uint64(p.Timestamp)
|
enc.Timestamp = hexutil.Uint64(p.Timestamp)
|
||||||
|
|
@ -28,6 +29,7 @@ func (p PayloadAttributes) MarshalJSON() ([]byte, error) {
|
||||||
enc.SuggestedFeeRecipient = p.SuggestedFeeRecipient
|
enc.SuggestedFeeRecipient = p.SuggestedFeeRecipient
|
||||||
enc.Withdrawals = p.Withdrawals
|
enc.Withdrawals = p.Withdrawals
|
||||||
enc.BeaconRoot = p.BeaconRoot
|
enc.BeaconRoot = p.BeaconRoot
|
||||||
|
enc.SlotNumber = (*hexutil.Uint64)(p.SlotNumber)
|
||||||
return json.Marshal(&enc)
|
return json.Marshal(&enc)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
@ -39,6 +41,7 @@ func (p *PayloadAttributes) UnmarshalJSON(input []byte) error {
|
||||||
SuggestedFeeRecipient *common.Address `json:"suggestedFeeRecipient" gencodec:"required"`
|
SuggestedFeeRecipient *common.Address `json:"suggestedFeeRecipient" gencodec:"required"`
|
||||||
Withdrawals []*types.Withdrawal `json:"withdrawals"`
|
Withdrawals []*types.Withdrawal `json:"withdrawals"`
|
||||||
BeaconRoot *common.Hash `json:"parentBeaconBlockRoot"`
|
BeaconRoot *common.Hash `json:"parentBeaconBlockRoot"`
|
||||||
|
SlotNumber *hexutil.Uint64 `json:"slotNumber"`
|
||||||
}
|
}
|
||||||
var dec PayloadAttributes
|
var dec PayloadAttributes
|
||||||
if err := json.Unmarshal(input, &dec); err != nil {
|
if err := json.Unmarshal(input, &dec); err != nil {
|
||||||
|
|
@ -62,5 +65,8 @@ func (p *PayloadAttributes) UnmarshalJSON(input []byte) error {
|
||||||
if dec.BeaconRoot != nil {
|
if dec.BeaconRoot != nil {
|
||||||
p.BeaconRoot = dec.BeaconRoot
|
p.BeaconRoot = dec.BeaconRoot
|
||||||
}
|
}
|
||||||
|
if dec.SlotNumber != nil {
|
||||||
|
p.SlotNumber = (*uint64)(dec.SlotNumber)
|
||||||
|
}
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
|
||||||
|
|
@ -34,6 +34,7 @@ func (e ExecutableData) MarshalJSON() ([]byte, error) {
|
||||||
Withdrawals []*types.Withdrawal `json:"withdrawals"`
|
Withdrawals []*types.Withdrawal `json:"withdrawals"`
|
||||||
BlobGasUsed *hexutil.Uint64 `json:"blobGasUsed"`
|
BlobGasUsed *hexutil.Uint64 `json:"blobGasUsed"`
|
||||||
ExcessBlobGas *hexutil.Uint64 `json:"excessBlobGas"`
|
ExcessBlobGas *hexutil.Uint64 `json:"excessBlobGas"`
|
||||||
|
SlotNumber *hexutil.Uint64 `json:"slotNumber"`
|
||||||
}
|
}
|
||||||
var enc ExecutableData
|
var enc ExecutableData
|
||||||
enc.ParentHash = e.ParentHash
|
enc.ParentHash = e.ParentHash
|
||||||
|
|
@ -58,6 +59,7 @@ func (e ExecutableData) MarshalJSON() ([]byte, error) {
|
||||||
enc.Withdrawals = e.Withdrawals
|
enc.Withdrawals = e.Withdrawals
|
||||||
enc.BlobGasUsed = (*hexutil.Uint64)(e.BlobGasUsed)
|
enc.BlobGasUsed = (*hexutil.Uint64)(e.BlobGasUsed)
|
||||||
enc.ExcessBlobGas = (*hexutil.Uint64)(e.ExcessBlobGas)
|
enc.ExcessBlobGas = (*hexutil.Uint64)(e.ExcessBlobGas)
|
||||||
|
enc.SlotNumber = (*hexutil.Uint64)(e.SlotNumber)
|
||||||
return json.Marshal(&enc)
|
return json.Marshal(&enc)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
@ -81,6 +83,7 @@ func (e *ExecutableData) UnmarshalJSON(input []byte) error {
|
||||||
Withdrawals []*types.Withdrawal `json:"withdrawals"`
|
Withdrawals []*types.Withdrawal `json:"withdrawals"`
|
||||||
BlobGasUsed *hexutil.Uint64 `json:"blobGasUsed"`
|
BlobGasUsed *hexutil.Uint64 `json:"blobGasUsed"`
|
||||||
ExcessBlobGas *hexutil.Uint64 `json:"excessBlobGas"`
|
ExcessBlobGas *hexutil.Uint64 `json:"excessBlobGas"`
|
||||||
|
SlotNumber *hexutil.Uint64 `json:"slotNumber"`
|
||||||
}
|
}
|
||||||
var dec ExecutableData
|
var dec ExecutableData
|
||||||
if err := json.Unmarshal(input, &dec); err != nil {
|
if err := json.Unmarshal(input, &dec); err != nil {
|
||||||
|
|
@ -154,5 +157,8 @@ func (e *ExecutableData) UnmarshalJSON(input []byte) error {
|
||||||
if dec.ExcessBlobGas != nil {
|
if dec.ExcessBlobGas != nil {
|
||||||
e.ExcessBlobGas = (*uint64)(dec.ExcessBlobGas)
|
e.ExcessBlobGas = (*uint64)(dec.ExcessBlobGas)
|
||||||
}
|
}
|
||||||
|
if dec.SlotNumber != nil {
|
||||||
|
e.SlotNumber = (*uint64)(dec.SlotNumber)
|
||||||
|
}
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
|
||||||
|
|
@ -50,6 +50,13 @@ var (
|
||||||
// ExecutionPayloadV3 has the syntax of ExecutionPayloadV2 and appends the new
|
// ExecutionPayloadV3 has the syntax of ExecutionPayloadV2 and appends the new
|
||||||
// fields: blobGasUsed and excessBlobGas.
|
// fields: blobGasUsed and excessBlobGas.
|
||||||
PayloadV3 PayloadVersion = 0x3
|
PayloadV3 PayloadVersion = 0x3
|
||||||
|
|
||||||
|
// PayloadV4 is the identifier of ExecutionPayloadV4 introduced in amsterdam fork.
|
||||||
|
//
|
||||||
|
// https://github.com/ethereum/execution-apis/blob/main/src/engine/amsterdam.md#executionpayloadv4
|
||||||
|
// ExecutionPayloadV4 has the syntax of ExecutionPayloadV3 and appends the new
|
||||||
|
// field slotNumber.
|
||||||
|
PayloadV4 PayloadVersion = 0x4
|
||||||
)
|
)
|
||||||
|
|
||||||
//go:generate go run github.com/fjl/gencodec -type PayloadAttributes -field-override payloadAttributesMarshaling -out gen_blockparams.go
|
//go:generate go run github.com/fjl/gencodec -type PayloadAttributes -field-override payloadAttributesMarshaling -out gen_blockparams.go
|
||||||
|
|
@ -62,11 +69,13 @@ type PayloadAttributes struct {
|
||||||
SuggestedFeeRecipient common.Address `json:"suggestedFeeRecipient" gencodec:"required"`
|
SuggestedFeeRecipient common.Address `json:"suggestedFeeRecipient" gencodec:"required"`
|
||||||
Withdrawals []*types.Withdrawal `json:"withdrawals"`
|
Withdrawals []*types.Withdrawal `json:"withdrawals"`
|
||||||
BeaconRoot *common.Hash `json:"parentBeaconBlockRoot"`
|
BeaconRoot *common.Hash `json:"parentBeaconBlockRoot"`
|
||||||
|
SlotNumber *uint64 `json:"slotNumber"`
|
||||||
}
|
}
|
||||||
|
|
||||||
// JSON type overrides for PayloadAttributes.
|
// JSON type overrides for PayloadAttributes.
|
||||||
type payloadAttributesMarshaling struct {
|
type payloadAttributesMarshaling struct {
|
||||||
Timestamp hexutil.Uint64
|
Timestamp hexutil.Uint64
|
||||||
|
SlotNumber *hexutil.Uint64
|
||||||
}
|
}
|
||||||
|
|
||||||
//go:generate go run github.com/fjl/gencodec -type ExecutableData -field-override executableDataMarshaling -out gen_ed.go
|
//go:generate go run github.com/fjl/gencodec -type ExecutableData -field-override executableDataMarshaling -out gen_ed.go
|
||||||
|
|
@ -90,6 +99,7 @@ type ExecutableData struct {
|
||||||
Withdrawals []*types.Withdrawal `json:"withdrawals"`
|
Withdrawals []*types.Withdrawal `json:"withdrawals"`
|
||||||
BlobGasUsed *uint64 `json:"blobGasUsed"`
|
BlobGasUsed *uint64 `json:"blobGasUsed"`
|
||||||
ExcessBlobGas *uint64 `json:"excessBlobGas"`
|
ExcessBlobGas *uint64 `json:"excessBlobGas"`
|
||||||
|
SlotNumber *uint64 `json:"slotNumber"`
|
||||||
}
|
}
|
||||||
|
|
||||||
// JSON type overrides for executableData.
|
// JSON type overrides for executableData.
|
||||||
|
|
@ -104,6 +114,7 @@ type executableDataMarshaling struct {
|
||||||
Transactions []hexutil.Bytes
|
Transactions []hexutil.Bytes
|
||||||
BlobGasUsed *hexutil.Uint64
|
BlobGasUsed *hexutil.Uint64
|
||||||
ExcessBlobGas *hexutil.Uint64
|
ExcessBlobGas *hexutil.Uint64
|
||||||
|
SlotNumber *hexutil.Uint64
|
||||||
}
|
}
|
||||||
|
|
||||||
// StatelessPayloadStatusV1 is the result of a stateless payload execution.
|
// StatelessPayloadStatusV1 is the result of a stateless payload execution.
|
||||||
|
|
@ -213,7 +224,7 @@ func encodeTransactions(txs []*types.Transaction) [][]byte {
|
||||||
return enc
|
return enc
|
||||||
}
|
}
|
||||||
|
|
||||||
func decodeTransactions(enc [][]byte) ([]*types.Transaction, error) {
|
func DecodeTransactions(enc [][]byte) ([]*types.Transaction, error) {
|
||||||
var txs = make([]*types.Transaction, len(enc))
|
var txs = make([]*types.Transaction, len(enc))
|
||||||
for i, encTx := range enc {
|
for i, encTx := range enc {
|
||||||
var tx types.Transaction
|
var tx types.Transaction
|
||||||
|
|
@ -251,7 +262,7 @@ func ExecutableDataToBlock(data ExecutableData, versionedHashes []common.Hash, b
|
||||||
// for stateless execution, so it skips checking if the executable data hashes to
|
// for stateless execution, so it skips checking if the executable data hashes to
|
||||||
// the requested hash (stateless has to *compute* the root hash, it's not given).
|
// the requested hash (stateless has to *compute* the root hash, it's not given).
|
||||||
func ExecutableDataToBlockNoHash(data ExecutableData, versionedHashes []common.Hash, beaconRoot *common.Hash, requests [][]byte) (*types.Block, error) {
|
func ExecutableDataToBlockNoHash(data ExecutableData, versionedHashes []common.Hash, beaconRoot *common.Hash, requests [][]byte) (*types.Block, error) {
|
||||||
txs, err := decodeTransactions(data.Transactions)
|
txs, err := DecodeTransactions(data.Transactions)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
|
|
@ -313,6 +324,7 @@ func ExecutableDataToBlockNoHash(data ExecutableData, versionedHashes []common.H
|
||||||
BlobGasUsed: data.BlobGasUsed,
|
BlobGasUsed: data.BlobGasUsed,
|
||||||
ParentBeaconRoot: beaconRoot,
|
ParentBeaconRoot: beaconRoot,
|
||||||
RequestsHash: requestsHash,
|
RequestsHash: requestsHash,
|
||||||
|
SlotNumber: data.SlotNumber,
|
||||||
}
|
}
|
||||||
return types.NewBlockWithHeader(header).
|
return types.NewBlockWithHeader(header).
|
||||||
WithBody(types.Body{Transactions: txs, Uncles: nil, Withdrawals: data.Withdrawals}),
|
WithBody(types.Body{Transactions: txs, Uncles: nil, Withdrawals: data.Withdrawals}),
|
||||||
|
|
@ -340,6 +352,7 @@ func BlockToExecutableData(block *types.Block, fees *big.Int, sidecars []*types.
|
||||||
Withdrawals: block.Withdrawals(),
|
Withdrawals: block.Withdrawals(),
|
||||||
BlobGasUsed: block.BlobGasUsed(),
|
BlobGasUsed: block.BlobGasUsed(),
|
||||||
ExcessBlobGas: block.ExcessBlobGas(),
|
ExcessBlobGas: block.ExcessBlobGas(),
|
||||||
|
SlotNumber: block.SlotNumber(),
|
||||||
}
|
}
|
||||||
|
|
||||||
// Add blobs.
|
// Add blobs.
|
||||||
|
|
|
||||||
|
|
@ -5,81 +5,102 @@
|
||||||
# https://github.com/ethereum/execution-spec-tests/releases/download/v5.1.0
|
# https://github.com/ethereum/execution-spec-tests/releases/download/v5.1.0
|
||||||
a3192784375acec7eaec492799d5c5d0c47a2909a3cc40178898e4ecd20cc416 fixtures_develop.tar.gz
|
a3192784375acec7eaec492799d5c5d0c47a2909a3cc40178898e4ecd20cc416 fixtures_develop.tar.gz
|
||||||
|
|
||||||
# version:golang 1.25.1
|
# version:golang 1.25.7
|
||||||
# https://go.dev/dl/
|
# https://go.dev/dl/
|
||||||
d010c109cee94d80efe681eab46bdea491ac906bf46583c32e9f0dbb0bd1a594 go1.25.1.src.tar.gz
|
178f2832820274b43e177d32f06a3ebb0129e427dd20a5e4c88df2c1763cf10a go1.25.7.src.tar.gz
|
||||||
1d622468f767a1b9fe1e1e67bd6ce6744d04e0c68712adc689748bbeccb126bb go1.25.1.darwin-amd64.tar.gz
|
81bf2a1f20633f62d55d826d82dde3b0570cf1408a91e15781b266037299285b go1.25.7.aix-ppc64.tar.gz
|
||||||
68deebb214f39d542e518ebb0598a406ab1b5a22bba8ec9ade9f55fb4dd94a6c go1.25.1.darwin-arm64.tar.gz
|
bf5050a2152f4053837b886e8d9640c829dbacbc3370f913351eb0904cb706f5 go1.25.7.darwin-amd64.tar.gz
|
||||||
d03cdcbc9bd8baf5cf028de390478e9e2b3e4d0afe5a6582dedc19bfe6a263b2 go1.25.1.linux-386.tar.gz
|
ff18369ffad05c57d5bed888b660b31385f3c913670a83ef557cdfd98ea9ae1b go1.25.7.darwin-arm64.tar.gz
|
||||||
7716a0d940a0f6ae8e1f3b3f4f36299dc53e31b16840dbd171254312c41ca12e go1.25.1.linux-amd64.tar.gz
|
c5dccd7f192dd7b305dc209fb316ac1917776d74bd8e4d532ef2772f305bf42a go1.25.7.dragonfly-amd64.tar.gz
|
||||||
65a3e34fb2126f55b34e1edfc709121660e1be2dee6bdf405fc399a63a95a87d go1.25.1.linux-arm64.tar.gz
|
a2de97c8ac74bf64b0ae73fe9d379e61af530e061bc7f8f825044172ffe61a8b go1.25.7.freebsd-386.tar.gz
|
||||||
eb949be683e82a99e9861dafd7057e31ea40b161eae6c4cd18fdc0e8c4ae6225 go1.25.1.linux-armv6l.tar.gz
|
055f9e138787dcafa81eb0314c8ff70c6dd0f6dba1e8a6957fef5d5efd1ab8fd go1.25.7.freebsd-amd64.tar.gz
|
||||||
be13d5479b8c75438f2efcaa8c191fba3af684b3228abc9c99c7aa8502f34424 go1.25.1.windows-386.zip
|
60e7f7a7c990f0b9539ac8ed668155746997d404643a4eecd47b3dee1b7e710b go1.25.7.freebsd-arm.tar.gz
|
||||||
4a974de310e7ee1d523d2fcedb114ba5fa75408c98eb3652023e55ccf3fa7cab go1.25.1.windows-amd64.zip
|
631e03d5fd4c526e2f499154d8c6bf4cb081afb2fff171c428722afc9539d53a go1.25.7.freebsd-arm64.tar.gz
|
||||||
45ab4290adbd6ee9e7f18f0d57eaa9008fdbef590882778ed93eac3c8cca06c5 go1.25.1.aix-ppc64.tar.gz
|
8a264fd685823808140672812e3ad9c43f6ad59444c0dc14cdd3a1351839ddd5 go1.25.7.freebsd-riscv64.tar.gz
|
||||||
2e3c1549bed3124763774d648f291ac42611232f48320ebbd23517c909c09b81 go1.25.1.dragonfly-amd64.tar.gz
|
57c672447d906a1bcab98f2b11492d54521a791aacbb4994a25169e59cbe289a go1.25.7.illumos-amd64.tar.gz
|
||||||
dc0198dd4ec520e13f26798def8750544edf6448d8e9c43fd2a814e4885932af go1.25.1.freebsd-386.tar.gz
|
2866517e9ca81e6a2e85a930e9b11bc8a05cfeb2fc6dc6cb2765e7fb3c14b715 go1.25.7.linux-386.tar.gz
|
||||||
c4f1a7e7b258406e6f3b677ecdbd97bbb23ff9c0d44be4eb238a07d360f69ac8 go1.25.1.freebsd-amd64.tar.gz
|
12e6d6a191091ae27dc31f6efc630e3a3b8ba409baf3573d955b196fdf086005 go1.25.7.linux-amd64.tar.gz
|
||||||
7772fc5ff71ed39297ec0c1599fc54e399642c9b848eac989601040923b0de9c go1.25.1.freebsd-arm.tar.gz
|
ba611a53534135a81067240eff9508cd7e256c560edd5d8c2fef54f083c07129 go1.25.7.linux-arm64.tar.gz
|
||||||
5bb011d5d5b6218b12189f07aa0be618ab2002662fff1ca40afba7389735c207 go1.25.1.freebsd-arm64.tar.gz
|
1ba07e0eb86b839e72467f4b5c7a5597d07f30bcf5563c951410454f7cda5266 go1.25.7.linux-armv6l.tar.gz
|
||||||
ccac716240cb049bebfafcb7eebc3758512178a4c51fc26da9cc032035d850c8 go1.25.1.freebsd-riscv64.tar.gz
|
775753fc5952a334c415f08768df2f0b73a3228a16e8f5f63d545daacb4e3357 go1.25.7.linux-loong64.tar.gz
|
||||||
cc53910ffb9fcfdd988a9fa25b5423bae1cfa01b19616be646700e1f5453b466 go1.25.1.illumos-amd64.tar.gz
|
1a023bb367c5fbb4c637a2f6dc23ff17c6591ad929ce16ea88c74d857153b307 go1.25.7.linux-mips.tar.gz
|
||||||
efe809f923bcedab44bf7be2b3af8d182b512b1bf9c07d302e0c45d26c8f56f3 go1.25.1.linux-loong64.tar.gz
|
a8e97223d8aa6fdfd45f132a4784d2f536bbac5f3d63a24b63d33b6bfe1549af go1.25.7.linux-mips64.tar.gz
|
||||||
c0de33679f6ed68991dc42dc4a602e74a666e3e166c1748ee1b5d1a7ea2ffbb2 go1.25.1.linux-mips.tar.gz
|
eb9edb6223330d5e20275667c65dea076b064c08e595fe4eba5d7d6055cfaccf go1.25.7.linux-mips64le.tar.gz
|
||||||
c270f7b0c0bdfbcd54fef4481227c40d41bb518f9ae38ee930870f04a0a6a589 go1.25.1.linux-mips64.tar.gz
|
9c1e693552a5f9bb9e0012d1c5e01456ecefbc59bef53a77305222ce10aba368 go1.25.7.linux-mipsle.tar.gz
|
||||||
80be871ba9c944f34d1868cdf5047e1cf2e1289fe08cdb90e2453d2f0d6965ae go1.25.1.linux-mips64le.tar.gz
|
28a788798e7329acbbc0ac2caa5e4368b1e5ede646cc24429c991214cfb45c63 go1.25.7.linux-ppc64.tar.gz
|
||||||
9f09defa9bb22ebf2cde76162f40958564e57ce5c2b3649bc063bebcbc9294c1 go1.25.1.linux-mipsle.tar.gz
|
42124c0edc92464e2b37b2d7fcd3658f0c47ebd6a098732415a522be8cb88e3f go1.25.7.linux-ppc64le.tar.gz
|
||||||
2c76b7d278c1d43ad19d478ad3f0f05e7b782b64b90870701b314fa48b5f43c6 go1.25.1.linux-ppc64.tar.gz
|
88d59c6893c8425875d6eef8e3434bc2fa2552e5ad4c058c6cd8cd710a0301c8 go1.25.7.linux-riscv64.tar.gz
|
||||||
8b0c8d3ee5b1b5c28b6bd63dc4438792012e01d03b4bf7a61d985c87edab7d1f go1.25.1.linux-ppc64le.tar.gz
|
c6b77facf666dc68195ecab05dbf0ebb4e755b2a8b7734c759880557f1c29b0c go1.25.7.linux-s390x.tar.gz
|
||||||
22fe934a9d0c9c57275716c55b92d46ebd887cec3177c9140705efa9f84ba1e2 go1.25.1.linux-riscv64.tar.gz
|
f14c184d9ade0ee04c7735d4071257b90896ecbde1b32adae84135f055e6399b go1.25.7.netbsd-386.tar.gz
|
||||||
9cfe517ba423f59f3738ca5c3d907c103253cffbbcc2987142f79c5de8c1bf93 go1.25.1.linux-s390x.tar.gz
|
7e7389e404dca1088c31f0fc07f1dd60891d7182bcd621469c14f7e79eceb3ff go1.25.7.netbsd-amd64.tar.gz
|
||||||
6af8a08353e76205d5b743dd7a3f0126684f96f62be0a31b75daf9837e512c46 go1.25.1.netbsd-386.tar.gz
|
70388bb3ef2f03dbf1357e9056bd09034a67e018262557354f8cf549766b3f9d go1.25.7.netbsd-arm.tar.gz
|
||||||
e5d534ff362edb1bd8c8e10892b6a027c4c1482454245d1529167676498684c7 go1.25.1.netbsd-amd64.tar.gz
|
8c1cda9d25bfc9b18d24d5f95fc23949dd3ff99fa408a6cfa40e2cf12b07e362 go1.25.7.netbsd-arm64.tar.gz
|
||||||
88bcf39254fdcea6a199c1c27d787831b652427ce60851ae9e41a3d7eb477f45 go1.25.1.netbsd-arm.tar.gz
|
42f0d1bfbe39b8401cccb84dd66b30795b97bfc9620dfdc17c5cd4fcf6495cb0 go1.25.7.openbsd-386.tar.gz
|
||||||
d7c2eabe1d04ee47bcaea2816fdd90dbd25d90d4dfa756faa9786c788e4f3a4e go1.25.1.netbsd-arm64.tar.gz
|
e514879c0a28bc32123cd52c4c093de912477fe83f36a6d07517d066ef55391a go1.25.7.openbsd-amd64.tar.gz
|
||||||
14a2845977eb4dde11d929858c437a043467c427db87899935e90cee04a38d72 go1.25.1.openbsd-386.tar.gz
|
8cd22530695a0218232bf7efea8f162df1697a3106942ac4129b8c3de39ce4ef go1.25.7.openbsd-arm.tar.gz
|
||||||
d27ac54b38a13a09c81e67c82ac70d387037341c85c3399291c73e13e83fdd8c go1.25.1.openbsd-amd64.tar.gz
|
938720f6ebc0d1c53d7840321d3a31f29fd02496e84a6538f442a9311dc1cc9a go1.25.7.openbsd-arm64.tar.gz
|
||||||
0f4ab5f02500afa4befd51fed1e8b45e4d07ca050f641cc3acc76eaa4027b2c3 go1.25.1.openbsd-arm.tar.gz
|
a4c378b73b98f89a3596c2ef51aabbb28783d9ca29f7e317d8ca07939660ce6f go1.25.7.openbsd-ppc64.tar.gz
|
||||||
d46c3bd156843656f7f3cb0dec27ea51cd926ec3f7b80744bf8156e67c1c812f go1.25.1.openbsd-arm64.tar.gz
|
937b58734fbeaa8c7941a0e4285e7e84b7885396e8d11c23f9ab1a8ff10ff20e go1.25.7.openbsd-riscv64.tar.gz
|
||||||
c550514c67f22e409be10e40eace761e2e43069f4ef086ae6e60aac736c2b679 go1.25.1.openbsd-ppc64.tar.gz
|
61a093c8c5244916f25740316386bb9f141545dcf01b06a79d1c78ece488403e go1.25.7.plan9-386.tar.gz
|
||||||
8a09a8714a2556eb13fc1f10b7ce2553fcea4971e3330fc3be0efd24aab45734 go1.25.1.openbsd-riscv64.tar.gz
|
7fc8f6689c9de8ccb7689d2278035fa83c2d601409101840df6ddfe09ba58699 go1.25.7.plan9-amd64.tar.gz
|
||||||
b0e1fefaf0c7abd71f139a54eee9767944aff5f0bc9d69c968234804884e552f go1.25.1.plan9-386.tar.gz
|
9661dff8eaeeb62f1c3aadbc5ff189a2e6744e1ec885e32dbcb438f58a34def5 go1.25.7.plan9-arm.tar.gz
|
||||||
e94732c94f149690aa0ab11c26090577211b4a988137cb2c03ec0b54e750402e go1.25.1.plan9-amd64.tar.gz
|
28ecba0e1d7950c8b29a4a04962dd49c3bf5221f55a44f17d98f369f82859cf4 go1.25.7.solaris-amd64.tar.gz
|
||||||
7eb80e9de1e817d9089a54e8c7c5c8d8ed9e5fb4d4a012fc0f18fc422a484f0c go1.25.1.plan9-arm.tar.gz
|
baa6b488291801642fa620026169e38bec2da2ac187cd3ae2145721cf826bbc3 go1.25.7.windows-386.zip
|
||||||
1261dfad7c4953c0ab90381bc1242dc54e394db7485c59349428d532b2273343 go1.25.1.solaris-amd64.tar.gz
|
c75e5f4ff62d085cc0017be3ad19d5536f46825fa05db06ec468941f847e3228 go1.25.7.windows-amd64.zip
|
||||||
04bc3c078e9e904c4d58d6ac2532a5bdd402bd36a9ff0b5949b3c5e6006a05ee go1.25.1.windows-arm64.zip
|
807033f85931bc4a589ca8497535dcbeb1f30d506e47fa200f5f04c4a71c3d9f go1.25.7.windows-arm64.zip
|
||||||
|
|
||||||
# version:golangci 2.4.0
|
# version:golangci 2.10.1
|
||||||
# https://github.com/golangci/golangci-lint/releases/
|
# https://github.com/golangci/golangci-lint/releases/
|
||||||
# https://github.com/golangci/golangci-lint/releases/download/v2.4.0/
|
# https://github.com/golangci/golangci-lint/releases/download/v2.10.1
|
||||||
7904ce63f79db44934939cf7a063086ea0ea98e9b19eba0a9d52ccdd0d21951c golangci-lint-2.4.0-darwin-amd64.tar.gz
|
66fb0da81b8033b477f97eea420d4b46b230ca172b8bb87c6610109f3772b6b6 golangci-lint-2.10.1-darwin-amd64.tar.gz
|
||||||
cd4dd53fa09b6646baff5fd22b8c64d91db02c21c7496df27992d75d34feec59 golangci-lint-2.4.0-darwin-arm64.tar.gz
|
03bfadf67e52b441b7ec21305e501c717df93c959836d66c7f97312654acb297 golangci-lint-2.10.1-darwin-arm64.tar.gz
|
||||||
d58f426ebe14cc257e81562b4bf37a488ffb4ffbbb3ec73041eb3b38bb25c0e1 golangci-lint-2.4.0-freebsd-386.tar.gz
|
c9a44658ccc8f7b8dbbd4ae6020ba91c1a5d3987f4d91ced0f7d2bea013e57ca golangci-lint-2.10.1-freebsd-386.tar.gz
|
||||||
6ec4a6177fc6c0dd541fbcb3a7612845266d020d35cc6fa92959220cdf64ca39 golangci-lint-2.4.0-freebsd-amd64.tar.gz
|
a513c5cb4e0f5bd5767001af9d5e97e7868cfc2d9c46739a4df93e713cfb24af golangci-lint-2.10.1-freebsd-amd64.tar.gz
|
||||||
4d473e3e71c01feaa915a0604fb35758b41284fb976cdeac3f842118d9ee7e17 golangci-lint-2.4.0-freebsd-armv6.tar.gz
|
2ef38eefc4b5cee2febacb75a30579526e5656c16338a921d80e59a8e87d4425 golangci-lint-2.10.1-freebsd-arm64.tar.gz
|
||||||
58727746c6530801a3f9a702a5945556a5eb7e88809222536dd9f9d54cafaeff golangci-lint-2.4.0-freebsd-armv7.tar.gz
|
8fea6766318b4829e766bbe325f10191d75297dcc44ae35bf374816037878e38 golangci-lint-2.10.1-freebsd-armv6.tar.gz
|
||||||
fbf28c662760e24c32f82f8d16dffdb4a82de7726a52ba1fad94f890c22997ea golangci-lint-2.4.0-illumos-amd64.tar.gz
|
30b629870574d6254f3e8804e5a74b34f98e1263c9d55465830d739c88b862ed golangci-lint-2.10.1-freebsd-armv7.tar.gz
|
||||||
a15a000a8981ef665e971e0f67e2acda9066a9e37a59344393b7351d8fb49c81 golangci-lint-2.4.0-linux-386.tar.gz
|
c0db839f866ce80b1b6c96167aa101cfe50d9c936f42d942a3c1cbdc1801af68 golangci-lint-2.10.1-illumos-amd64.tar.gz
|
||||||
fae792524c04424c0ac369f5b8076f04b45cf29fc945a370e55d369a8dc11840 golangci-lint-2.4.0-linux-amd64.tar.gz
|
280eb56636e9175f671cd7b755d7d67f628ae2ed00a164d1e443c43c112034e5 golangci-lint-2.10.1-linux-386.deb
|
||||||
70ac11f55b80ec78fd3a879249cc9255121b8dfd7f7ed4fc46ed137f4abf17e7 golangci-lint-2.4.0-linux-arm64.tar.gz
|
065a7d99da61dc7dfbfef2e2d7053dd3fa6672598f2747117aa4bb5f45e7df7f golangci-lint-2.10.1-linux-386.rpm
|
||||||
4acdc40e5cebe99e4e7ced358a05b2e71789f409b41cb4f39bbb86ccfa14b1dc golangci-lint-2.4.0-linux-armv6.tar.gz
|
a55918c03bb413b2662287653ab2ae2fef4e37428b247dad6348724adde9d770 golangci-lint-2.10.1-linux-386.tar.gz
|
||||||
2a68749568fa22b4a97cb88dbea655595563c795076536aa6c087f7968784bf3 golangci-lint-2.4.0-linux-armv7.tar.gz
|
8aa9b3aa14f39745eeb7fc7ff50bcac683e785397d1e4bc9afd2184b12c4ce86 golangci-lint-2.10.1-linux-amd64.deb
|
||||||
9e3369afb023711036dcb0b4f45c9fe2792af962fa1df050c9f6ac101a6c5d73 golangci-lint-2.4.0-linux-loong64.tar.gz
|
62a111688e9e305032334a2cbc84f4d971b64bb3bffc99d3f80081d57fb25e32 golangci-lint-2.10.1-linux-amd64.rpm
|
||||||
bb9143d6329be2c4dbfffef9564078e7da7d88e7dde6c829b6263d98e072229e golangci-lint-2.4.0-linux-mips64.tar.gz
|
dfa775874cf0561b404a02a8f4481fc69b28091da95aa697259820d429b09c99 golangci-lint-2.10.1-linux-amd64.tar.gz
|
||||||
5ad1765b40d56cd04d4afd805b3ba6f4bfd9b36181da93c31e9b17e483d8608d golangci-lint-2.4.0-linux-mips64le.tar.gz
|
b3f36937e8ea1660739dc0f5c892ea59c9c21ed4e75a91a25957c561f7f79a55 golangci-lint-2.10.1-linux-arm64.deb
|
||||||
918936fb9c0d5ba96bef03cf4348b03938634cfcced49be1e9bb29cb5094fa73 golangci-lint-2.4.0-linux-ppc64le.tar.gz
|
36d50314d53683b1f1a2a6cedfb5a9468451b481c64ab9e97a8e843ea088074d golangci-lint-2.10.1-linux-arm64.rpm
|
||||||
f7474c638e1fb67ebbdc654b55ca0125377ea0bc88e8fee8d964a4f24eacf828 golangci-lint-2.4.0-linux-riscv64.tar.gz
|
6652b42ae02915eb2f9cb2a2e0cac99514c8eded8388d88ae3e06e1a52c00de8 golangci-lint-2.10.1-linux-arm64.tar.gz
|
||||||
b617a9543997c8bfceaffa88a75d4e595030c6add69fba800c1e4d8f5fe253dd golangci-lint-2.4.0-linux-s390x.tar.gz
|
a32d8d318e803496812dd3461f250e52ccc7f53c47b95ce404a9cf55778ceb6a golangci-lint-2.10.1-linux-armv6.deb
|
||||||
7db027b03a9ba328f795215b04f594036837bc7dd0dd7cd16776b02a6167981c golangci-lint-2.4.0-netbsd-386.tar.gz
|
41d065f4c8ea165a1531abea644988ee2e973e4f0b49f9725ed3b979dac45112 golangci-lint-2.10.1-linux-armv6.rpm
|
||||||
52d8f9393f4313df0a62b752c37775e3af0b818e43e8dd28954351542d7c60bc golangci-lint-2.4.0-netbsd-amd64.tar.gz
|
59159a4df03aabbde69d15c7b7b3df143363cbb41f4bd4b200caffb8e34fb734 golangci-lint-2.10.1-linux-armv6.tar.gz
|
||||||
5c0086027fb5a4af3829e530c8115db4b35d11afe1914322eef528eb8cd38c69 golangci-lint-2.4.0-netbsd-arm64.tar.gz
|
b2e8ec0e050a1e2251dfe1561434999d202f5a3f9fa47ce94378b0fd1662ea5a golangci-lint-2.10.1-linux-armv7.deb
|
||||||
6b779d6ed1aed87cefe195cc11759902b97a76551b593312c6833f2635a3488f golangci-lint-2.4.0-netbsd-armv6.tar.gz
|
28c9331429a497da27e9c77846063bd0e8275e878ffedb4eb9e9f21d24771cc0 golangci-lint-2.10.1-linux-armv7.rpm
|
||||||
f00d1f4b7ec3468a0f9fffd0d9ea036248b029b7621cbc9a59c449ef94356d09 golangci-lint-2.4.0-netbsd-armv7.tar.gz
|
818f33e95b273e3769284b25563b51ef6a294e9e25acf140fda5830c075a1a59 golangci-lint-2.10.1-linux-armv7.tar.gz
|
||||||
3ce671b0b42b58e35066493aab75a7e2826c9e079988f1ba5d814a4029faaf87 golangci-lint-2.4.0-windows-386.zip
|
6b6b85ed4b7c27f51097dd681523000409dde835e86e6e314e87be4bb013e2ab golangci-lint-2.10.1-linux-loong64.deb
|
||||||
003112f7a56746feaabf20b744054bf9acdf900c9e77176383623c4b1d76aaa9 golangci-lint-2.4.0-windows-amd64.zip
|
94050a0cf06169e2ae44afb307dcaafa7d7c3b38c0c23b5652cf9cb60f0c337f golangci-lint-2.10.1-linux-loong64.rpm
|
||||||
dc0c2092af5d47fc2cd31a1dfe7b4c7e765fab22de98bd21ef2ffcc53ad9f54f golangci-lint-2.4.0-windows-arm64.zip
|
25820300fccb8c961c1cdcb1f77928040c079e04c43a3a5ceb34b1cb4a1c5c8d golangci-lint-2.10.1-linux-loong64.tar.gz
|
||||||
0263d23e20a260cb1592d35e12a388f99efe2c51b3611fdc66fbd9db1fce664d golangci-lint-2.4.0-windows-armv6.zip
|
98bf39d10139fdcaa37f94950e9bbb8888660ae468847ae0bf1cb5bf67c1f68b golangci-lint-2.10.1-linux-mips64.deb
|
||||||
9403c03bf648e6313036e0273149d44bad1b9ad53889b6d00e4ccb842ba3c058 golangci-lint-2.4.0-windows-armv7.zip
|
df3ce5f03808dcceaa8b683d1d06e95c885f09b59dc8e15deb840fbe2b3e3299 golangci-lint-2.10.1-linux-mips64.rpm
|
||||||
|
972508dda523067e6e6a1c8e6609d63bc7c4153819c11b947d439235cf17bac2 golangci-lint-2.10.1-linux-mips64.tar.gz
|
||||||
|
1d37f2919e183b5bf8b1777ed8c4b163d3b491d0158355a7999d647655cbbeb6 golangci-lint-2.10.1-linux-mips64le.deb
|
||||||
|
e341d031002cd09a416329ed40f674231051a38544b8f94deb2d1708ce1f4a6f golangci-lint-2.10.1-linux-mips64le.rpm
|
||||||
|
393560122b9cb5538df0c357d30eb27b6ee563533fbb9b138c8db4fd264002af golangci-lint-2.10.1-linux-mips64le.tar.gz
|
||||||
|
21ca46b6a96442e8957677a3ca059c6b93674a68a01b1c71f4e5df0ea2e96d19 golangci-lint-2.10.1-linux-ppc64le.deb
|
||||||
|
57fe0cbca0a9bbdf1547c5e8aa7d278e6896b438d72a541bae6bc62c38b43d1e golangci-lint-2.10.1-linux-ppc64le.rpm
|
||||||
|
e2883db9fa51584e5e203c64456f29993550a7faadc84e3faccdb48f0669992e golangci-lint-2.10.1-linux-ppc64le.tar.gz
|
||||||
|
aa6da0e98ab0ba3bb7582e112174c349907d5edfeff90a551dca3c6eecf92fc0 golangci-lint-2.10.1-linux-riscv64.deb
|
||||||
|
3c68d76cd884a7aad206223a980b9c20bb9ea74b560fa27ed02baf2389189234 golangci-lint-2.10.1-linux-riscv64.rpm
|
||||||
|
3bca11bfac4197205639cbd4676a5415054e629ac6c12ea10fcbe33ef852d9c3 golangci-lint-2.10.1-linux-riscv64.tar.gz
|
||||||
|
0c6aed2ce49db2586adbac72c80d871f06feb1caf4c0763a5ca98fec809a8f0b golangci-lint-2.10.1-linux-s390x.deb
|
||||||
|
16c285adfe1061d69dd8e503be69f87c7202857c6f4add74ac02e3571158fbec golangci-lint-2.10.1-linux-s390x.rpm
|
||||||
|
21011ad368eb04f024201b832095c6b5f96d0888de194cca5bfe4d9307d6364b golangci-lint-2.10.1-linux-s390x.tar.gz
|
||||||
|
7b5191e77a70485918712e31ed55159956323e4911bab1b67569c9d86e1b75eb golangci-lint-2.10.1-netbsd-386.tar.gz
|
||||||
|
07801fd38d293ebad10826f8285525a39ea91ce5ddad77d05bfa90bda9c884a9 golangci-lint-2.10.1-netbsd-amd64.tar.gz
|
||||||
|
7e7219d71c1bf33b98c328c93dc0560706dd896a1c43c44696e5222fc9d7446e golangci-lint-2.10.1-netbsd-arm64.tar.gz
|
||||||
|
92fbc90b9eec0e572269b0f5492a2895c426b086a68372fde49b7e4d4020863e golangci-lint-2.10.1-netbsd-armv6.tar.gz
|
||||||
|
f67b3ae1f47caeefa507a4ebb0c8336958a19011fe48766443212030f75d004b golangci-lint-2.10.1-netbsd-armv7.tar.gz
|
||||||
|
a40bc091c10cea84eaee1a90b84b65f5e8652113b0a600bb099e4e4d9d7caddb golangci-lint-2.10.1-windows-386.zip
|
||||||
|
c60c87695e79db8e320f0e5be885059859de52bb5ee5f11be5577828570bc2a3 golangci-lint-2.10.1-windows-amd64.zip
|
||||||
|
636ab790c8dcea8034aa34aba6031ca3893d68f7eda000460ab534341fadbab1 golangci-lint-2.10.1-windows-arm64.zip
|
||||||
|
|
||||||
# This is the builder on PPA that will build Go itself (inception-y), don't modify!
|
# This is the builder on PPA that will build Go itself (inception-y), don't modify!
|
||||||
#
|
#
|
||||||
|
|
|
||||||
|
|
@ -168,8 +168,6 @@ var (
|
||||||
"focal", // 20.04, EOL: 04/2030
|
"focal", // 20.04, EOL: 04/2030
|
||||||
"jammy", // 22.04, EOL: 04/2032
|
"jammy", // 22.04, EOL: 04/2032
|
||||||
"noble", // 24.04, EOL: 04/2034
|
"noble", // 24.04, EOL: 04/2034
|
||||||
"oracular", // 24.10, EOL: 07/2025
|
|
||||||
"plucky", // 25.04, EOL: 01/2026
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// This is where the tests should be unpacked.
|
// This is where the tests should be unpacked.
|
||||||
|
|
|
||||||
|
|
@ -155,7 +155,7 @@ func (c *Conn) ReadEth() (any, error) {
|
||||||
var msg any
|
var msg any
|
||||||
switch int(code) {
|
switch int(code) {
|
||||||
case eth.StatusMsg:
|
case eth.StatusMsg:
|
||||||
msg = new(eth.StatusPacket69)
|
msg = new(eth.StatusPacket)
|
||||||
case eth.GetBlockHeadersMsg:
|
case eth.GetBlockHeadersMsg:
|
||||||
msg = new(eth.GetBlockHeadersPacket)
|
msg = new(eth.GetBlockHeadersPacket)
|
||||||
case eth.BlockHeadersMsg:
|
case eth.BlockHeadersMsg:
|
||||||
|
|
@ -164,10 +164,6 @@ func (c *Conn) ReadEth() (any, error) {
|
||||||
msg = new(eth.GetBlockBodiesPacket)
|
msg = new(eth.GetBlockBodiesPacket)
|
||||||
case eth.BlockBodiesMsg:
|
case eth.BlockBodiesMsg:
|
||||||
msg = new(eth.BlockBodiesPacket)
|
msg = new(eth.BlockBodiesPacket)
|
||||||
case eth.NewBlockMsg:
|
|
||||||
msg = new(eth.NewBlockPacket)
|
|
||||||
case eth.NewBlockHashesMsg:
|
|
||||||
msg = new(eth.NewBlockHashesPacket)
|
|
||||||
case eth.TransactionsMsg:
|
case eth.TransactionsMsg:
|
||||||
msg = new(eth.TransactionsPacket)
|
msg = new(eth.TransactionsPacket)
|
||||||
case eth.NewPooledTransactionHashesMsg:
|
case eth.NewPooledTransactionHashesMsg:
|
||||||
|
|
@ -229,7 +225,7 @@ func (c *Conn) ReadSnap() (any, error) {
|
||||||
}
|
}
|
||||||
|
|
||||||
// dialAndPeer creates a peer connection and runs the handshake.
|
// dialAndPeer creates a peer connection and runs the handshake.
|
||||||
func (s *Suite) dialAndPeer(status *eth.StatusPacket69) (*Conn, error) {
|
func (s *Suite) dialAndPeer(status *eth.StatusPacket) (*Conn, error) {
|
||||||
c, err := s.dial()
|
c, err := s.dial()
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
|
|
@ -242,7 +238,7 @@ func (s *Suite) dialAndPeer(status *eth.StatusPacket69) (*Conn, error) {
|
||||||
|
|
||||||
// peer performs both the protocol handshake and the status message
|
// peer performs both the protocol handshake and the status message
|
||||||
// exchange with the node in order to peer with it.
|
// exchange with the node in order to peer with it.
|
||||||
func (c *Conn) peer(chain *Chain, status *eth.StatusPacket69) error {
|
func (c *Conn) peer(chain *Chain, status *eth.StatusPacket) error {
|
||||||
if err := c.handshake(); err != nil {
|
if err := c.handshake(); err != nil {
|
||||||
return fmt.Errorf("handshake failed: %v", err)
|
return fmt.Errorf("handshake failed: %v", err)
|
||||||
}
|
}
|
||||||
|
|
@ -315,7 +311,7 @@ func (c *Conn) negotiateEthProtocol(caps []p2p.Cap) {
|
||||||
}
|
}
|
||||||
|
|
||||||
// statusExchange performs a `Status` message exchange with the given node.
|
// statusExchange performs a `Status` message exchange with the given node.
|
||||||
func (c *Conn) statusExchange(chain *Chain, status *eth.StatusPacket69) error {
|
func (c *Conn) statusExchange(chain *Chain, status *eth.StatusPacket) error {
|
||||||
loop:
|
loop:
|
||||||
for {
|
for {
|
||||||
code, data, err := c.Read()
|
code, data, err := c.Read()
|
||||||
|
|
@ -324,7 +320,7 @@ loop:
|
||||||
}
|
}
|
||||||
switch code {
|
switch code {
|
||||||
case eth.StatusMsg + protoOffset(ethProto):
|
case eth.StatusMsg + protoOffset(ethProto):
|
||||||
msg := new(eth.StatusPacket69)
|
msg := new(eth.StatusPacket)
|
||||||
if err := rlp.DecodeBytes(data, &msg); err != nil {
|
if err := rlp.DecodeBytes(data, &msg); err != nil {
|
||||||
return fmt.Errorf("error decoding status packet: %w", err)
|
return fmt.Errorf("error decoding status packet: %w", err)
|
||||||
}
|
}
|
||||||
|
|
@ -363,7 +359,7 @@ loop:
|
||||||
}
|
}
|
||||||
if status == nil {
|
if status == nil {
|
||||||
// default status message
|
// default status message
|
||||||
status = ð.StatusPacket69{
|
status = ð.StatusPacket{
|
||||||
ProtocolVersion: uint32(c.negotiatedProtoVersion),
|
ProtocolVersion: uint32(c.negotiatedProtoVersion),
|
||||||
NetworkID: chain.config.ChainID.Uint64(),
|
NetworkID: chain.config.ChainID.Uint64(),
|
||||||
Genesis: chain.blocks[0].Hash(),
|
Genesis: chain.blocks[0].Hash(),
|
||||||
|
|
|
||||||
|
|
@ -447,7 +447,7 @@ func (s *Suite) TestGetReceipts(t *utesting.T) {
|
||||||
t.Fatalf("could not write to connection: %v", err)
|
t.Fatalf("could not write to connection: %v", err)
|
||||||
}
|
}
|
||||||
// Wait for response.
|
// Wait for response.
|
||||||
resp := new(eth.ReceiptsPacket[*eth.ReceiptList69])
|
resp := new(eth.ReceiptsPacket)
|
||||||
if err := conn.ReadMsg(ethProto, eth.ReceiptsMsg, &resp); err != nil {
|
if err := conn.ReadMsg(ethProto, eth.ReceiptsMsg, &resp); err != nil {
|
||||||
t.Fatalf("error reading block bodies msg: %v", err)
|
t.Fatalf("error reading block bodies msg: %v", err)
|
||||||
}
|
}
|
||||||
|
|
|
||||||
|
|
@ -52,7 +52,7 @@ func (s *Suite) AllTests() []utesting.Test {
|
||||||
{Name: "Ping", Fn: s.TestPing},
|
{Name: "Ping", Fn: s.TestPing},
|
||||||
{Name: "PingLargeRequestID", Fn: s.TestPingLargeRequestID},
|
{Name: "PingLargeRequestID", Fn: s.TestPingLargeRequestID},
|
||||||
{Name: "PingMultiIP", Fn: s.TestPingMultiIP},
|
{Name: "PingMultiIP", Fn: s.TestPingMultiIP},
|
||||||
{Name: "PingHandshakeInterrupted", Fn: s.TestPingHandshakeInterrupted},
|
{Name: "HandshakeResend", Fn: s.TestHandshakeResend},
|
||||||
{Name: "TalkRequest", Fn: s.TestTalkRequest},
|
{Name: "TalkRequest", Fn: s.TestTalkRequest},
|
||||||
{Name: "FindnodeZeroDistance", Fn: s.TestFindnodeZeroDistance},
|
{Name: "FindnodeZeroDistance", Fn: s.TestFindnodeZeroDistance},
|
||||||
{Name: "FindnodeResults", Fn: s.TestFindnodeResults},
|
{Name: "FindnodeResults", Fn: s.TestFindnodeResults},
|
||||||
|
|
@ -158,22 +158,20 @@ the attempt from a different IP.`)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// TestPingHandshakeInterrupted starts a handshake, but doesn't finish it and sends a second ordinary message
|
// TestHandshakeResend starts a handshake, but doesn't finish it and sends a second ordinary message
|
||||||
// packet instead of a handshake message packet. The remote node should respond with
|
// packet instead of a handshake message packet. The remote node should repeat the previous WHOAREYOU
|
||||||
// another WHOAREYOU challenge for the second packet.
|
// challenge for the first PING.
|
||||||
func (s *Suite) TestPingHandshakeInterrupted(t *utesting.T) {
|
func (s *Suite) TestHandshakeResend(t *utesting.T) {
|
||||||
t.Log(`TestPingHandshakeInterrupted starts a handshake, but doesn't finish it and sends a second ordinary message
|
|
||||||
packet instead of a handshake message packet. The remote node should respond with
|
|
||||||
another WHOAREYOU challenge for the second packet.`)
|
|
||||||
|
|
||||||
conn, l1 := s.listen1(t)
|
conn, l1 := s.listen1(t)
|
||||||
defer conn.close()
|
defer conn.close()
|
||||||
|
|
||||||
// First PING triggers challenge.
|
// First PING triggers challenge.
|
||||||
ping := &v5wire.Ping{ReqID: conn.nextReqID()}
|
ping := &v5wire.Ping{ReqID: conn.nextReqID()}
|
||||||
conn.write(l1, ping, nil)
|
conn.write(l1, ping, nil)
|
||||||
|
var challenge1 *v5wire.Whoareyou
|
||||||
switch resp := conn.read(l1).(type) {
|
switch resp := conn.read(l1).(type) {
|
||||||
case *v5wire.Whoareyou:
|
case *v5wire.Whoareyou:
|
||||||
|
challenge1 = resp
|
||||||
t.Logf("got WHOAREYOU for PING")
|
t.Logf("got WHOAREYOU for PING")
|
||||||
default:
|
default:
|
||||||
t.Fatal("expected WHOAREYOU, got", resp)
|
t.Fatal("expected WHOAREYOU, got", resp)
|
||||||
|
|
@ -181,9 +179,16 @@ another WHOAREYOU challenge for the second packet.`)
|
||||||
|
|
||||||
// Send second PING.
|
// Send second PING.
|
||||||
ping2 := &v5wire.Ping{ReqID: conn.nextReqID()}
|
ping2 := &v5wire.Ping{ReqID: conn.nextReqID()}
|
||||||
switch resp := conn.reqresp(l1, ping2).(type) {
|
conn.write(l1, ping2, nil)
|
||||||
case *v5wire.Pong:
|
switch resp := conn.read(l1).(type) {
|
||||||
checkPong(t, resp, ping2, l1)
|
case *v5wire.Whoareyou:
|
||||||
|
if resp.Nonce != challenge1.Nonce {
|
||||||
|
t.Fatalf("wrong nonce %x in WHOAREYOU (want %x)", resp.Nonce[:], challenge1.Nonce[:])
|
||||||
|
}
|
||||||
|
if !bytes.Equal(resp.ChallengeData, challenge1.ChallengeData) {
|
||||||
|
t.Fatalf("wrong ChallengeData in resent WHOAREYOU (want %x)", resp.ChallengeData, challenge1.ChallengeData)
|
||||||
|
}
|
||||||
|
resp.Node = conn.remote
|
||||||
default:
|
default:
|
||||||
t.Fatal("expected WHOAREYOU, got", resp)
|
t.Fatal("expected WHOAREYOU, got", resp)
|
||||||
}
|
}
|
||||||
|
|
|
||||||
|
|
@ -56,6 +56,7 @@ type header struct {
|
||||||
BlobGasUsed *uint64 `json:"blobGasUsed" rlp:"optional"`
|
BlobGasUsed *uint64 `json:"blobGasUsed" rlp:"optional"`
|
||||||
ExcessBlobGas *uint64 `json:"excessBlobGas" rlp:"optional"`
|
ExcessBlobGas *uint64 `json:"excessBlobGas" rlp:"optional"`
|
||||||
ParentBeaconBlockRoot *common.Hash `json:"parentBeaconBlockRoot" rlp:"optional"`
|
ParentBeaconBlockRoot *common.Hash `json:"parentBeaconBlockRoot" rlp:"optional"`
|
||||||
|
SlotNumber *uint64 `json:"slotNumber" rlp:"optional"`
|
||||||
}
|
}
|
||||||
|
|
||||||
type headerMarshaling struct {
|
type headerMarshaling struct {
|
||||||
|
|
@ -68,6 +69,7 @@ type headerMarshaling struct {
|
||||||
BaseFee *math.HexOrDecimal256
|
BaseFee *math.HexOrDecimal256
|
||||||
BlobGasUsed *math.HexOrDecimal64
|
BlobGasUsed *math.HexOrDecimal64
|
||||||
ExcessBlobGas *math.HexOrDecimal64
|
ExcessBlobGas *math.HexOrDecimal64
|
||||||
|
SlotNumber *math.HexOrDecimal64
|
||||||
}
|
}
|
||||||
|
|
||||||
type bbInput struct {
|
type bbInput struct {
|
||||||
|
|
@ -136,6 +138,7 @@ func (i *bbInput) ToBlock() *types.Block {
|
||||||
BlobGasUsed: i.Header.BlobGasUsed,
|
BlobGasUsed: i.Header.BlobGasUsed,
|
||||||
ExcessBlobGas: i.Header.ExcessBlobGas,
|
ExcessBlobGas: i.Header.ExcessBlobGas,
|
||||||
ParentBeaconRoot: i.Header.ParentBeaconBlockRoot,
|
ParentBeaconRoot: i.Header.ParentBeaconBlockRoot,
|
||||||
|
SlotNumber: i.Header.SlotNumber,
|
||||||
}
|
}
|
||||||
|
|
||||||
// Fill optional values.
|
// Fill optional values.
|
||||||
|
|
|
||||||
|
|
@ -102,6 +102,7 @@ type stEnv struct {
|
||||||
ParentExcessBlobGas *uint64 `json:"parentExcessBlobGas,omitempty"`
|
ParentExcessBlobGas *uint64 `json:"parentExcessBlobGas,omitempty"`
|
||||||
ParentBlobGasUsed *uint64 `json:"parentBlobGasUsed,omitempty"`
|
ParentBlobGasUsed *uint64 `json:"parentBlobGasUsed,omitempty"`
|
||||||
ParentBeaconBlockRoot *common.Hash `json:"parentBeaconBlockRoot"`
|
ParentBeaconBlockRoot *common.Hash `json:"parentBeaconBlockRoot"`
|
||||||
|
SlotNumber *uint64 `json:"slotNumber"`
|
||||||
}
|
}
|
||||||
|
|
||||||
type stEnvMarshaling struct {
|
type stEnvMarshaling struct {
|
||||||
|
|
@ -120,6 +121,7 @@ type stEnvMarshaling struct {
|
||||||
ExcessBlobGas *math.HexOrDecimal64
|
ExcessBlobGas *math.HexOrDecimal64
|
||||||
ParentExcessBlobGas *math.HexOrDecimal64
|
ParentExcessBlobGas *math.HexOrDecimal64
|
||||||
ParentBlobGasUsed *math.HexOrDecimal64
|
ParentBlobGasUsed *math.HexOrDecimal64
|
||||||
|
SlotNumber *math.HexOrDecimal64
|
||||||
}
|
}
|
||||||
|
|
||||||
type rejectedTx struct {
|
type rejectedTx struct {
|
||||||
|
|
@ -147,15 +149,13 @@ func (pre *Prestate) Apply(vmConfig vm.Config, chainConfig *params.ChainConfig,
|
||||||
isEIP4762 = chainConfig.IsVerkle(big.NewInt(int64(pre.Env.Number)), pre.Env.Timestamp)
|
isEIP4762 = chainConfig.IsVerkle(big.NewInt(int64(pre.Env.Number)), pre.Env.Timestamp)
|
||||||
statedb = MakePreState(rawdb.NewMemoryDatabase(), pre.Pre, isEIP4762)
|
statedb = MakePreState(rawdb.NewMemoryDatabase(), pre.Pre, isEIP4762)
|
||||||
signer = types.MakeSigner(chainConfig, new(big.Int).SetUint64(pre.Env.Number), pre.Env.Timestamp)
|
signer = types.MakeSigner(chainConfig, new(big.Int).SetUint64(pre.Env.Number), pre.Env.Timestamp)
|
||||||
gaspool = new(core.GasPool)
|
gaspool = core.NewGasPool(pre.Env.GasLimit)
|
||||||
blockHash = common.Hash{0x13, 0x37}
|
blockHash = common.Hash{0x13, 0x37}
|
||||||
rejectedTxs []*rejectedTx
|
rejectedTxs []*rejectedTx
|
||||||
includedTxs types.Transactions
|
includedTxs types.Transactions
|
||||||
gasUsed = uint64(0)
|
|
||||||
blobGasUsed = uint64(0)
|
blobGasUsed = uint64(0)
|
||||||
receipts = make(types.Receipts, 0)
|
receipts = make(types.Receipts, 0)
|
||||||
)
|
)
|
||||||
gaspool.AddGas(pre.Env.GasLimit)
|
|
||||||
vmContext := vm.BlockContext{
|
vmContext := vm.BlockContext{
|
||||||
CanTransfer: core.CanTransfer,
|
CanTransfer: core.CanTransfer,
|
||||||
Transfer: core.Transfer,
|
Transfer: core.Transfer,
|
||||||
|
|
@ -195,6 +195,7 @@ func (pre *Prestate) Apply(vmConfig vm.Config, chainConfig *params.ChainConfig,
|
||||||
ExcessBlobGas: pre.Env.ParentExcessBlobGas,
|
ExcessBlobGas: pre.Env.ParentExcessBlobGas,
|
||||||
BlobGasUsed: pre.Env.ParentBlobGasUsed,
|
BlobGasUsed: pre.Env.ParentBlobGasUsed,
|
||||||
BaseFee: pre.Env.ParentBaseFee,
|
BaseFee: pre.Env.ParentBaseFee,
|
||||||
|
SlotNumber: pre.Env.SlotNumber,
|
||||||
}
|
}
|
||||||
header := &types.Header{
|
header := &types.Header{
|
||||||
Time: pre.Env.Timestamp,
|
Time: pre.Env.Timestamp,
|
||||||
|
|
@ -255,16 +256,19 @@ func (pre *Prestate) Apply(vmConfig vm.Config, chainConfig *params.ChainConfig,
|
||||||
statedb.SetTxContext(tx.Hash(), len(receipts))
|
statedb.SetTxContext(tx.Hash(), len(receipts))
|
||||||
var (
|
var (
|
||||||
snapshot = statedb.Snapshot()
|
snapshot = statedb.Snapshot()
|
||||||
prevGas = gaspool.Gas()
|
gp = gaspool.Snapshot()
|
||||||
)
|
)
|
||||||
receipt, err := core.ApplyTransactionWithEVM(msg, gaspool, statedb, vmContext.BlockNumber, blockHash, pre.Env.Timestamp, tx, &gasUsed, evm)
|
receipt, err := core.ApplyTransactionWithEVM(msg, gaspool, statedb, vmContext.BlockNumber, blockHash, pre.Env.Timestamp, tx, evm)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
statedb.RevertToSnapshot(snapshot)
|
statedb.RevertToSnapshot(snapshot)
|
||||||
log.Info("rejected tx", "index", i, "hash", tx.Hash(), "from", msg.From, "error", err)
|
log.Info("rejected tx", "index", i, "hash", tx.Hash(), "from", msg.From, "error", err)
|
||||||
rejectedTxs = append(rejectedTxs, &rejectedTx{i, err.Error()})
|
rejectedTxs = append(rejectedTxs, &rejectedTx{i, err.Error()})
|
||||||
gaspool.SetGas(prevGas)
|
gaspool.Set(gp)
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
|
if receipt.Logs == nil {
|
||||||
|
receipt.Logs = []*types.Log{}
|
||||||
|
}
|
||||||
includedTxs = append(includedTxs, tx)
|
includedTxs = append(includedTxs, tx)
|
||||||
if hashError != nil {
|
if hashError != nil {
|
||||||
return nil, nil, nil, NewError(ErrorMissingBlockhash, hashError)
|
return nil, nil, nil, NewError(ErrorMissingBlockhash, hashError)
|
||||||
|
|
@ -346,7 +350,7 @@ func (pre *Prestate) Apply(vmConfig vm.Config, chainConfig *params.ChainConfig,
|
||||||
Receipts: receipts,
|
Receipts: receipts,
|
||||||
Rejected: rejectedTxs,
|
Rejected: rejectedTxs,
|
||||||
Difficulty: (*math.HexOrDecimal256)(vmContext.Difficulty),
|
Difficulty: (*math.HexOrDecimal256)(vmContext.Difficulty),
|
||||||
GasUsed: (math.HexOrDecimal64)(gasUsed),
|
GasUsed: (math.HexOrDecimal64)(gaspool.Used()),
|
||||||
BaseFee: (*math.HexOrDecimal256)(vmContext.BaseFee),
|
BaseFee: (*math.HexOrDecimal256)(vmContext.BaseFee),
|
||||||
}
|
}
|
||||||
if pre.Env.Withdrawals != nil {
|
if pre.Env.Withdrawals != nil {
|
||||||
|
|
|
||||||
|
|
@ -56,27 +56,35 @@ func (l *fileWritingTracer) Write(p []byte) (n int, err error) {
|
||||||
return n, nil
|
return n, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// newFileWriter creates a set of hooks which wraps inner hooks (typically a logger),
|
// newFileWriter creates a tracer which wraps inner hooks (typically a logger),
|
||||||
// and writes the output to a file, one file per transaction.
|
// and writes the output to a file, one file per transaction.
|
||||||
func newFileWriter(baseDir string, innerFn func(out io.Writer) *tracing.Hooks) *tracing.Hooks {
|
func newFileWriter(baseDir string, innerFn func(out io.Writer) *tracing.Hooks) *tracers.Tracer {
|
||||||
t := &fileWritingTracer{
|
t := &fileWritingTracer{
|
||||||
baseDir: baseDir,
|
baseDir: baseDir,
|
||||||
suffix: "jsonl",
|
suffix: "jsonl",
|
||||||
}
|
}
|
||||||
t.inner = innerFn(t) // instantiate the inner tracer
|
t.inner = innerFn(t) // instantiate the inner tracer
|
||||||
return t.hooks()
|
return &tracers.Tracer{
|
||||||
|
Hooks: t.hooks(),
|
||||||
|
GetResult: func() (json.RawMessage, error) { return json.RawMessage("{}"), nil },
|
||||||
|
Stop: func(err error) {},
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// newResultWriter creates a set of hooks wraps and invokes an underlying tracer,
|
// newResultWriter creates a tracer that wraps and invokes an underlying tracer,
|
||||||
// and writes the result (getResult-output) to file, one per transaction.
|
// and writes the result (getResult-output) to file, one per transaction.
|
||||||
func newResultWriter(baseDir string, tracer *tracers.Tracer) *tracing.Hooks {
|
func newResultWriter(baseDir string, tracer *tracers.Tracer) *tracers.Tracer {
|
||||||
t := &fileWritingTracer{
|
t := &fileWritingTracer{
|
||||||
baseDir: baseDir,
|
baseDir: baseDir,
|
||||||
getResult: tracer.GetResult,
|
getResult: tracer.GetResult,
|
||||||
inner: tracer.Hooks,
|
inner: tracer.Hooks,
|
||||||
suffix: "json",
|
suffix: "json",
|
||||||
}
|
}
|
||||||
return t.hooks()
|
return &tracers.Tracer{
|
||||||
|
Hooks: t.hooks(),
|
||||||
|
GetResult: func() (json.RawMessage, error) { return json.RawMessage("{}"), nil },
|
||||||
|
Stop: func(err error) {},
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// OnTxStart creates a new output-file specific for this transaction, and invokes
|
// OnTxStart creates a new output-file specific for this transaction, and invokes
|
||||||
|
|
|
||||||
|
|
@ -162,6 +162,11 @@ var (
|
||||||
strings.Join(vm.ActivateableEips(), ", ")),
|
strings.Join(vm.ActivateableEips(), ", ")),
|
||||||
Value: "GrayGlacier",
|
Value: "GrayGlacier",
|
||||||
}
|
}
|
||||||
|
OpcodeCountFlag = &cli.StringFlag{
|
||||||
|
Name: "opcode.count",
|
||||||
|
Usage: "If set, opcode execution counts will be written to this file (relative to output.basedir).",
|
||||||
|
Value: "",
|
||||||
|
}
|
||||||
VerbosityFlag = &cli.IntFlag{
|
VerbosityFlag = &cli.IntFlag{
|
||||||
Name: "verbosity",
|
Name: "verbosity",
|
||||||
Usage: "sets the verbosity level",
|
Usage: "sets the verbosity level",
|
||||||
|
|
|
||||||
|
|
@ -38,6 +38,7 @@ func (h header) MarshalJSON() ([]byte, error) {
|
||||||
BlobGasUsed *math.HexOrDecimal64 `json:"blobGasUsed" rlp:"optional"`
|
BlobGasUsed *math.HexOrDecimal64 `json:"blobGasUsed" rlp:"optional"`
|
||||||
ExcessBlobGas *math.HexOrDecimal64 `json:"excessBlobGas" rlp:"optional"`
|
ExcessBlobGas *math.HexOrDecimal64 `json:"excessBlobGas" rlp:"optional"`
|
||||||
ParentBeaconBlockRoot *common.Hash `json:"parentBeaconBlockRoot" rlp:"optional"`
|
ParentBeaconBlockRoot *common.Hash `json:"parentBeaconBlockRoot" rlp:"optional"`
|
||||||
|
SlotNumber *math.HexOrDecimal64 `json:"slotNumber" rlp:"optional"`
|
||||||
}
|
}
|
||||||
var enc header
|
var enc header
|
||||||
enc.ParentHash = h.ParentHash
|
enc.ParentHash = h.ParentHash
|
||||||
|
|
@ -60,6 +61,7 @@ func (h header) MarshalJSON() ([]byte, error) {
|
||||||
enc.BlobGasUsed = (*math.HexOrDecimal64)(h.BlobGasUsed)
|
enc.BlobGasUsed = (*math.HexOrDecimal64)(h.BlobGasUsed)
|
||||||
enc.ExcessBlobGas = (*math.HexOrDecimal64)(h.ExcessBlobGas)
|
enc.ExcessBlobGas = (*math.HexOrDecimal64)(h.ExcessBlobGas)
|
||||||
enc.ParentBeaconBlockRoot = h.ParentBeaconBlockRoot
|
enc.ParentBeaconBlockRoot = h.ParentBeaconBlockRoot
|
||||||
|
enc.SlotNumber = (*math.HexOrDecimal64)(h.SlotNumber)
|
||||||
return json.Marshal(&enc)
|
return json.Marshal(&enc)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
@ -86,6 +88,7 @@ func (h *header) UnmarshalJSON(input []byte) error {
|
||||||
BlobGasUsed *math.HexOrDecimal64 `json:"blobGasUsed" rlp:"optional"`
|
BlobGasUsed *math.HexOrDecimal64 `json:"blobGasUsed" rlp:"optional"`
|
||||||
ExcessBlobGas *math.HexOrDecimal64 `json:"excessBlobGas" rlp:"optional"`
|
ExcessBlobGas *math.HexOrDecimal64 `json:"excessBlobGas" rlp:"optional"`
|
||||||
ParentBeaconBlockRoot *common.Hash `json:"parentBeaconBlockRoot" rlp:"optional"`
|
ParentBeaconBlockRoot *common.Hash `json:"parentBeaconBlockRoot" rlp:"optional"`
|
||||||
|
SlotNumber *math.HexOrDecimal64 `json:"slotNumber" rlp:"optional"`
|
||||||
}
|
}
|
||||||
var dec header
|
var dec header
|
||||||
if err := json.Unmarshal(input, &dec); err != nil {
|
if err := json.Unmarshal(input, &dec); err != nil {
|
||||||
|
|
@ -155,5 +158,8 @@ func (h *header) UnmarshalJSON(input []byte) error {
|
||||||
if dec.ParentBeaconBlockRoot != nil {
|
if dec.ParentBeaconBlockRoot != nil {
|
||||||
h.ParentBeaconBlockRoot = dec.ParentBeaconBlockRoot
|
h.ParentBeaconBlockRoot = dec.ParentBeaconBlockRoot
|
||||||
}
|
}
|
||||||
|
if dec.SlotNumber != nil {
|
||||||
|
h.SlotNumber = (*uint64)(dec.SlotNumber)
|
||||||
|
}
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
|
||||||
|
|
@ -37,6 +37,7 @@ func (s stEnv) MarshalJSON() ([]byte, error) {
|
||||||
ParentExcessBlobGas *math.HexOrDecimal64 `json:"parentExcessBlobGas,omitempty"`
|
ParentExcessBlobGas *math.HexOrDecimal64 `json:"parentExcessBlobGas,omitempty"`
|
||||||
ParentBlobGasUsed *math.HexOrDecimal64 `json:"parentBlobGasUsed,omitempty"`
|
ParentBlobGasUsed *math.HexOrDecimal64 `json:"parentBlobGasUsed,omitempty"`
|
||||||
ParentBeaconBlockRoot *common.Hash `json:"parentBeaconBlockRoot"`
|
ParentBeaconBlockRoot *common.Hash `json:"parentBeaconBlockRoot"`
|
||||||
|
SlotNumber *math.HexOrDecimal64 `json:"slotNumber"`
|
||||||
}
|
}
|
||||||
var enc stEnv
|
var enc stEnv
|
||||||
enc.Coinbase = common.UnprefixedAddress(s.Coinbase)
|
enc.Coinbase = common.UnprefixedAddress(s.Coinbase)
|
||||||
|
|
@ -59,6 +60,7 @@ func (s stEnv) MarshalJSON() ([]byte, error) {
|
||||||
enc.ParentExcessBlobGas = (*math.HexOrDecimal64)(s.ParentExcessBlobGas)
|
enc.ParentExcessBlobGas = (*math.HexOrDecimal64)(s.ParentExcessBlobGas)
|
||||||
enc.ParentBlobGasUsed = (*math.HexOrDecimal64)(s.ParentBlobGasUsed)
|
enc.ParentBlobGasUsed = (*math.HexOrDecimal64)(s.ParentBlobGasUsed)
|
||||||
enc.ParentBeaconBlockRoot = s.ParentBeaconBlockRoot
|
enc.ParentBeaconBlockRoot = s.ParentBeaconBlockRoot
|
||||||
|
enc.SlotNumber = (*math.HexOrDecimal64)(s.SlotNumber)
|
||||||
return json.Marshal(&enc)
|
return json.Marshal(&enc)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
@ -85,6 +87,7 @@ func (s *stEnv) UnmarshalJSON(input []byte) error {
|
||||||
ParentExcessBlobGas *math.HexOrDecimal64 `json:"parentExcessBlobGas,omitempty"`
|
ParentExcessBlobGas *math.HexOrDecimal64 `json:"parentExcessBlobGas,omitempty"`
|
||||||
ParentBlobGasUsed *math.HexOrDecimal64 `json:"parentBlobGasUsed,omitempty"`
|
ParentBlobGasUsed *math.HexOrDecimal64 `json:"parentBlobGasUsed,omitempty"`
|
||||||
ParentBeaconBlockRoot *common.Hash `json:"parentBeaconBlockRoot"`
|
ParentBeaconBlockRoot *common.Hash `json:"parentBeaconBlockRoot"`
|
||||||
|
SlotNumber *math.HexOrDecimal64 `json:"slotNumber"`
|
||||||
}
|
}
|
||||||
var dec stEnv
|
var dec stEnv
|
||||||
if err := json.Unmarshal(input, &dec); err != nil {
|
if err := json.Unmarshal(input, &dec); err != nil {
|
||||||
|
|
@ -154,5 +157,8 @@ func (s *stEnv) UnmarshalJSON(input []byte) error {
|
||||||
if dec.ParentBeaconBlockRoot != nil {
|
if dec.ParentBeaconBlockRoot != nil {
|
||||||
s.ParentBeaconBlockRoot = dec.ParentBeaconBlockRoot
|
s.ParentBeaconBlockRoot = dec.ParentBeaconBlockRoot
|
||||||
}
|
}
|
||||||
|
if dec.SlotNumber != nil {
|
||||||
|
s.SlotNumber = (*uint64)(dec.SlotNumber)
|
||||||
|
}
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
|
||||||
|
|
@ -37,6 +37,7 @@ import (
|
||||||
"github.com/ethereum/go-ethereum/crypto"
|
"github.com/ethereum/go-ethereum/crypto"
|
||||||
"github.com/ethereum/go-ethereum/eth/tracers"
|
"github.com/ethereum/go-ethereum/eth/tracers"
|
||||||
"github.com/ethereum/go-ethereum/eth/tracers/logger"
|
"github.com/ethereum/go-ethereum/eth/tracers/logger"
|
||||||
|
"github.com/ethereum/go-ethereum/eth/tracers/native"
|
||||||
"github.com/ethereum/go-ethereum/log"
|
"github.com/ethereum/go-ethereum/log"
|
||||||
"github.com/ethereum/go-ethereum/params"
|
"github.com/ethereum/go-ethereum/params"
|
||||||
"github.com/ethereum/go-ethereum/tests"
|
"github.com/ethereum/go-ethereum/tests"
|
||||||
|
|
@ -167,14 +168,15 @@ func Transition(ctx *cli.Context) error {
|
||||||
}
|
}
|
||||||
|
|
||||||
// Configure tracer
|
// Configure tracer
|
||||||
|
var tracer *tracers.Tracer
|
||||||
if ctx.IsSet(TraceTracerFlag.Name) { // Custom tracing
|
if ctx.IsSet(TraceTracerFlag.Name) { // Custom tracing
|
||||||
config := json.RawMessage(ctx.String(TraceTracerConfigFlag.Name))
|
config := json.RawMessage(ctx.String(TraceTracerConfigFlag.Name))
|
||||||
tracer, err := tracers.DefaultDirectory.New(ctx.String(TraceTracerFlag.Name),
|
innerTracer, err := tracers.DefaultDirectory.New(ctx.String(TraceTracerFlag.Name),
|
||||||
nil, config, chainConfig)
|
nil, config, chainConfig)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return NewError(ErrorConfig, fmt.Errorf("failed instantiating tracer: %v", err))
|
return NewError(ErrorConfig, fmt.Errorf("failed instantiating tracer: %v", err))
|
||||||
}
|
}
|
||||||
vmConfig.Tracer = newResultWriter(baseDir, tracer)
|
tracer = newResultWriter(baseDir, innerTracer)
|
||||||
} else if ctx.Bool(TraceFlag.Name) { // JSON opcode tracing
|
} else if ctx.Bool(TraceFlag.Name) { // JSON opcode tracing
|
||||||
logConfig := &logger.Config{
|
logConfig := &logger.Config{
|
||||||
DisableStack: ctx.Bool(TraceDisableStackFlag.Name),
|
DisableStack: ctx.Bool(TraceDisableStackFlag.Name),
|
||||||
|
|
@ -182,20 +184,45 @@ func Transition(ctx *cli.Context) error {
|
||||||
EnableReturnData: ctx.Bool(TraceEnableReturnDataFlag.Name),
|
EnableReturnData: ctx.Bool(TraceEnableReturnDataFlag.Name),
|
||||||
}
|
}
|
||||||
if ctx.Bool(TraceEnableCallFramesFlag.Name) {
|
if ctx.Bool(TraceEnableCallFramesFlag.Name) {
|
||||||
vmConfig.Tracer = newFileWriter(baseDir, func(out io.Writer) *tracing.Hooks {
|
tracer = newFileWriter(baseDir, func(out io.Writer) *tracing.Hooks {
|
||||||
return logger.NewJSONLoggerWithCallFrames(logConfig, out)
|
return logger.NewJSONLoggerWithCallFrames(logConfig, out)
|
||||||
})
|
})
|
||||||
} else {
|
} else {
|
||||||
vmConfig.Tracer = newFileWriter(baseDir, func(out io.Writer) *tracing.Hooks {
|
tracer = newFileWriter(baseDir, func(out io.Writer) *tracing.Hooks {
|
||||||
return logger.NewJSONLogger(logConfig, out)
|
return logger.NewJSONLogger(logConfig, out)
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
// Configure opcode counter
|
||||||
|
var opcodeTracer *tracers.Tracer
|
||||||
|
if ctx.IsSet(OpcodeCountFlag.Name) && ctx.String(OpcodeCountFlag.Name) != "" {
|
||||||
|
opcodeTracer = native.NewOpcodeCounter()
|
||||||
|
if tracer != nil {
|
||||||
|
// If we have an existing tracer, multiplex with the opcode tracer
|
||||||
|
mux, _ := native.NewMuxTracer([]string{"trace", "opcode"}, []*tracers.Tracer{tracer, opcodeTracer})
|
||||||
|
vmConfig.Tracer = mux.Hooks
|
||||||
|
} else {
|
||||||
|
vmConfig.Tracer = opcodeTracer.Hooks
|
||||||
|
}
|
||||||
|
} else if tracer != nil {
|
||||||
|
vmConfig.Tracer = tracer.Hooks
|
||||||
|
}
|
||||||
// Run the test and aggregate the result
|
// Run the test and aggregate the result
|
||||||
s, result, body, err := prestate.Apply(vmConfig, chainConfig, txIt, ctx.Int64(RewardFlag.Name))
|
s, result, body, err := prestate.Apply(vmConfig, chainConfig, txIt, ctx.Int64(RewardFlag.Name))
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
// Write opcode counts if enabled
|
||||||
|
if opcodeTracer != nil {
|
||||||
|
fname := ctx.String(OpcodeCountFlag.Name)
|
||||||
|
result, err := opcodeTracer.GetResult()
|
||||||
|
if err != nil {
|
||||||
|
return NewError(ErrorJson, fmt.Errorf("failed getting opcode counts: %v", err))
|
||||||
|
}
|
||||||
|
if err := saveFile(baseDir, fname, result); err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
}
|
||||||
// Dump the execution result
|
// Dump the execution result
|
||||||
var (
|
var (
|
||||||
collector = make(Alloc)
|
collector = make(Alloc)
|
||||||
|
|
|
||||||
|
|
@ -161,6 +161,7 @@ var (
|
||||||
t8ntool.ForknameFlag,
|
t8ntool.ForknameFlag,
|
||||||
t8ntool.ChainIDFlag,
|
t8ntool.ChainIDFlag,
|
||||||
t8ntool.RewardFlag,
|
t8ntool.RewardFlag,
|
||||||
|
t8ntool.OpcodeCountFlag,
|
||||||
},
|
},
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
||||||
2
cmd/evm/testdata/1/exp.json
vendored
2
cmd/evm/testdata/1/exp.json
vendored
|
|
@ -24,7 +24,7 @@
|
||||||
"status": "0x1",
|
"status": "0x1",
|
||||||
"cumulativeGasUsed": "0x5208",
|
"cumulativeGasUsed": "0x5208",
|
||||||
"logsBloom": "0x00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000",
|
"logsBloom": "0x00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000",
|
||||||
"logs": null,
|
"logs": [],
|
||||||
"transactionHash": "0x0557bacce3375c98d806609b8d5043072f0b6a8bae45ae5a67a00d3a1a18d673",
|
"transactionHash": "0x0557bacce3375c98d806609b8d5043072f0b6a8bae45ae5a67a00d3a1a18d673",
|
||||||
"contractAddress": "0x0000000000000000000000000000000000000000",
|
"contractAddress": "0x0000000000000000000000000000000000000000",
|
||||||
"gasUsed": "0x5208",
|
"gasUsed": "0x5208",
|
||||||
|
|
|
||||||
4
cmd/evm/testdata/13/exp2.json
vendored
4
cmd/evm/testdata/13/exp2.json
vendored
|
|
@ -12,7 +12,7 @@
|
||||||
"status": "0x0",
|
"status": "0x0",
|
||||||
"cumulativeGasUsed": "0x84d0",
|
"cumulativeGasUsed": "0x84d0",
|
||||||
"logsBloom": "0x00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000",
|
"logsBloom": "0x00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000",
|
||||||
"logs": null,
|
"logs": [],
|
||||||
"transactionHash": "0xa98a24882ea90916c6a86da650fbc6b14238e46f0af04a131ce92be897507476",
|
"transactionHash": "0xa98a24882ea90916c6a86da650fbc6b14238e46f0af04a131ce92be897507476",
|
||||||
"contractAddress": "0x0000000000000000000000000000000000000000",
|
"contractAddress": "0x0000000000000000000000000000000000000000",
|
||||||
"gasUsed": "0x84d0",
|
"gasUsed": "0x84d0",
|
||||||
|
|
@ -27,7 +27,7 @@
|
||||||
"status": "0x0",
|
"status": "0x0",
|
||||||
"cumulativeGasUsed": "0x109a0",
|
"cumulativeGasUsed": "0x109a0",
|
||||||
"logsBloom": "0x00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000",
|
"logsBloom": "0x00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000",
|
||||||
"logs": null,
|
"logs": [],
|
||||||
"transactionHash": "0x36bad80acce7040c45fd32764b5c2b2d2e6f778669fb41791f73f546d56e739a",
|
"transactionHash": "0x36bad80acce7040c45fd32764b5c2b2d2e6f778669fb41791f73f546d56e739a",
|
||||||
"contractAddress": "0x0000000000000000000000000000000000000000",
|
"contractAddress": "0x0000000000000000000000000000000000000000",
|
||||||
"gasUsed": "0x84d0",
|
"gasUsed": "0x84d0",
|
||||||
|
|
|
||||||
2
cmd/evm/testdata/23/exp.json
vendored
2
cmd/evm/testdata/23/exp.json
vendored
|
|
@ -11,7 +11,7 @@
|
||||||
"status": "0x1",
|
"status": "0x1",
|
||||||
"cumulativeGasUsed": "0x520b",
|
"cumulativeGasUsed": "0x520b",
|
||||||
"logsBloom": "0x00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000",
|
"logsBloom": "0x00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000",
|
||||||
"logs": null,
|
"logs": [],
|
||||||
"transactionHash": "0x72fadbef39cd251a437eea619cfeda752271a5faaaa2147df012e112159ffb81",
|
"transactionHash": "0x72fadbef39cd251a437eea619cfeda752271a5faaaa2147df012e112159ffb81",
|
||||||
"contractAddress": "0x0000000000000000000000000000000000000000",
|
"contractAddress": "0x0000000000000000000000000000000000000000",
|
||||||
"gasUsed": "0x520b",
|
"gasUsed": "0x520b",
|
||||||
|
|
|
||||||
4
cmd/evm/testdata/24/exp.json
vendored
4
cmd/evm/testdata/24/exp.json
vendored
|
|
@ -27,7 +27,7 @@
|
||||||
"status": "0x1",
|
"status": "0x1",
|
||||||
"cumulativeGasUsed": "0xa861",
|
"cumulativeGasUsed": "0xa861",
|
||||||
"logsBloom": "0x00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000",
|
"logsBloom": "0x00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000",
|
||||||
"logs": null,
|
"logs": [],
|
||||||
"transactionHash": "0x92ea4a28224d033afb20e0cc2b290d4c7c2d61f6a4800a680e4e19ac962ee941",
|
"transactionHash": "0x92ea4a28224d033afb20e0cc2b290d4c7c2d61f6a4800a680e4e19ac962ee941",
|
||||||
"contractAddress": "0x0000000000000000000000000000000000000000",
|
"contractAddress": "0x0000000000000000000000000000000000000000",
|
||||||
"gasUsed": "0xa861",
|
"gasUsed": "0xa861",
|
||||||
|
|
@ -41,7 +41,7 @@
|
||||||
"status": "0x1",
|
"status": "0x1",
|
||||||
"cumulativeGasUsed": "0x10306",
|
"cumulativeGasUsed": "0x10306",
|
||||||
"logsBloom": "0x00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000",
|
"logsBloom": "0x00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000",
|
||||||
"logs": null,
|
"logs": [],
|
||||||
"transactionHash": "0x16b1d912f1d664f3f60f4e1b5f296f3c82a64a1a253117b4851d18bc03c4f1da",
|
"transactionHash": "0x16b1d912f1d664f3f60f4e1b5f296f3c82a64a1a253117b4851d18bc03c4f1da",
|
||||||
"contractAddress": "0x0000000000000000000000000000000000000000",
|
"contractAddress": "0x0000000000000000000000000000000000000000",
|
||||||
"gasUsed": "0x5aa5",
|
"gasUsed": "0x5aa5",
|
||||||
|
|
|
||||||
2
cmd/evm/testdata/25/exp.json
vendored
2
cmd/evm/testdata/25/exp.json
vendored
|
|
@ -23,7 +23,7 @@
|
||||||
"status": "0x1",
|
"status": "0x1",
|
||||||
"cumulativeGasUsed": "0x5208",
|
"cumulativeGasUsed": "0x5208",
|
||||||
"logsBloom": "0x00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000",
|
"logsBloom": "0x00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000",
|
||||||
"logs": null,
|
"logs": [],
|
||||||
"transactionHash": "0x92ea4a28224d033afb20e0cc2b290d4c7c2d61f6a4800a680e4e19ac962ee941",
|
"transactionHash": "0x92ea4a28224d033afb20e0cc2b290d4c7c2d61f6a4800a680e4e19ac962ee941",
|
||||||
"contractAddress": "0x0000000000000000000000000000000000000000",
|
"contractAddress": "0x0000000000000000000000000000000000000000",
|
||||||
"gasUsed": "0x5208",
|
"gasUsed": "0x5208",
|
||||||
|
|
|
||||||
2
cmd/evm/testdata/28/exp.json
vendored
2
cmd/evm/testdata/28/exp.json
vendored
|
|
@ -28,7 +28,7 @@
|
||||||
"status": "0x1",
|
"status": "0x1",
|
||||||
"cumulativeGasUsed": "0xa865",
|
"cumulativeGasUsed": "0xa865",
|
||||||
"logsBloom": "0x00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000",
|
"logsBloom": "0x00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000",
|
||||||
"logs": null,
|
"logs": [],
|
||||||
"transactionHash": "0x7508d7139d002a4b3a26a4f12dec0d87cb46075c78bf77a38b569a133b509262",
|
"transactionHash": "0x7508d7139d002a4b3a26a4f12dec0d87cb46075c78bf77a38b569a133b509262",
|
||||||
"contractAddress": "0x0000000000000000000000000000000000000000",
|
"contractAddress": "0x0000000000000000000000000000000000000000",
|
||||||
"gasUsed": "0xa865",
|
"gasUsed": "0xa865",
|
||||||
|
|
|
||||||
2
cmd/evm/testdata/29/exp.json
vendored
2
cmd/evm/testdata/29/exp.json
vendored
|
|
@ -26,7 +26,7 @@
|
||||||
"status": "0x1",
|
"status": "0x1",
|
||||||
"cumulativeGasUsed": "0x5208",
|
"cumulativeGasUsed": "0x5208",
|
||||||
"logsBloom": "0x00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000",
|
"logsBloom": "0x00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000",
|
||||||
"logs": null,
|
"logs": [],
|
||||||
"transactionHash": "0x84f70aba406a55628a0620f26d260f90aeb6ccc55fed6ec2ac13dd4f727032ed",
|
"transactionHash": "0x84f70aba406a55628a0620f26d260f90aeb6ccc55fed6ec2ac13dd4f727032ed",
|
||||||
"contractAddress": "0x0000000000000000000000000000000000000000",
|
"contractAddress": "0x0000000000000000000000000000000000000000",
|
||||||
"gasUsed": "0x5208",
|
"gasUsed": "0x5208",
|
||||||
|
|
|
||||||
2
cmd/evm/testdata/3/exp.json
vendored
2
cmd/evm/testdata/3/exp.json
vendored
|
|
@ -24,7 +24,7 @@
|
||||||
"status": "0x1",
|
"status": "0x1",
|
||||||
"cumulativeGasUsed": "0x521f",
|
"cumulativeGasUsed": "0x521f",
|
||||||
"logsBloom": "0x00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000",
|
"logsBloom": "0x00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000",
|
||||||
"logs": null,
|
"logs": [],
|
||||||
"transactionHash": "0x72fadbef39cd251a437eea619cfeda752271a5faaaa2147df012e112159ffb81",
|
"transactionHash": "0x72fadbef39cd251a437eea619cfeda752271a5faaaa2147df012e112159ffb81",
|
||||||
"contractAddress": "0x0000000000000000000000000000000000000000",
|
"contractAddress": "0x0000000000000000000000000000000000000000",
|
||||||
"gasUsed": "0x521f",
|
"gasUsed": "0x521f",
|
||||||
|
|
|
||||||
4
cmd/evm/testdata/30/exp.json
vendored
4
cmd/evm/testdata/30/exp.json
vendored
|
|
@ -25,7 +25,7 @@
|
||||||
"status": "0x1",
|
"status": "0x1",
|
||||||
"cumulativeGasUsed": "0x5208",
|
"cumulativeGasUsed": "0x5208",
|
||||||
"logsBloom": "0x00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000",
|
"logsBloom": "0x00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000",
|
||||||
"logs": null,
|
"logs": [],
|
||||||
"transactionHash": "0xa98a24882ea90916c6a86da650fbc6b14238e46f0af04a131ce92be897507476",
|
"transactionHash": "0xa98a24882ea90916c6a86da650fbc6b14238e46f0af04a131ce92be897507476",
|
||||||
"contractAddress": "0x0000000000000000000000000000000000000000",
|
"contractAddress": "0x0000000000000000000000000000000000000000",
|
||||||
"gasUsed": "0x5208",
|
"gasUsed": "0x5208",
|
||||||
|
|
@ -40,7 +40,7 @@
|
||||||
"status": "0x1",
|
"status": "0x1",
|
||||||
"cumulativeGasUsed": "0xa410",
|
"cumulativeGasUsed": "0xa410",
|
||||||
"logsBloom": "0x00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000",
|
"logsBloom": "0x00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000",
|
||||||
"logs": null,
|
"logs": [],
|
||||||
"transactionHash": "0x36bad80acce7040c45fd32764b5c2b2d2e6f778669fb41791f73f546d56e739a",
|
"transactionHash": "0x36bad80acce7040c45fd32764b5c2b2d2e6f778669fb41791f73f546d56e739a",
|
||||||
"contractAddress": "0x0000000000000000000000000000000000000000",
|
"contractAddress": "0x0000000000000000000000000000000000000000",
|
||||||
"gasUsed": "0x5208",
|
"gasUsed": "0x5208",
|
||||||
|
|
|
||||||
2
cmd/evm/testdata/33/exp.json
vendored
2
cmd/evm/testdata/33/exp.json
vendored
|
|
@ -44,7 +44,7 @@
|
||||||
"root": "0x",
|
"root": "0x",
|
||||||
"status": "0x1",
|
"status": "0x1",
|
||||||
"cumulativeGasUsed": "0x15fa9",
|
"cumulativeGasUsed": "0x15fa9",
|
||||||
"logsBloom": "0x00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000","logs": null,"transactionHash": "0x0417aab7c1d8a3989190c3167c132876ce9b8afd99262c5a0f9d06802de3d7ef",
|
"logsBloom": "0x00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000","logs": [],"transactionHash": "0x0417aab7c1d8a3989190c3167c132876ce9b8afd99262c5a0f9d06802de3d7ef",
|
||||||
"contractAddress": "0x0000000000000000000000000000000000000000",
|
"contractAddress": "0x0000000000000000000000000000000000000000",
|
||||||
"gasUsed": "0x15fa9",
|
"gasUsed": "0x15fa9",
|
||||||
"effectiveGasPrice": null,
|
"effectiveGasPrice": null,
|
||||||
|
|
|
||||||
|
|
@ -111,6 +111,7 @@ if one is set. Otherwise it prints the genesis from the datadir.`,
|
||||||
utils.MetricsInfluxDBUsernameFlag,
|
utils.MetricsInfluxDBUsernameFlag,
|
||||||
utils.MetricsInfluxDBPasswordFlag,
|
utils.MetricsInfluxDBPasswordFlag,
|
||||||
utils.MetricsInfluxDBTagsFlag,
|
utils.MetricsInfluxDBTagsFlag,
|
||||||
|
utils.MetricsInfluxDBIntervalFlag,
|
||||||
utils.MetricsInfluxDBTokenFlag,
|
utils.MetricsInfluxDBTokenFlag,
|
||||||
utils.MetricsInfluxDBBucketFlag,
|
utils.MetricsInfluxDBBucketFlag,
|
||||||
utils.MetricsInfluxDBOrganizationFlag,
|
utils.MetricsInfluxDBOrganizationFlag,
|
||||||
|
|
|
||||||
|
|
@ -377,6 +377,9 @@ func applyMetricConfig(ctx *cli.Context, cfg *gethConfig) {
|
||||||
if ctx.IsSet(utils.MetricsInfluxDBTagsFlag.Name) {
|
if ctx.IsSet(utils.MetricsInfluxDBTagsFlag.Name) {
|
||||||
cfg.Metrics.InfluxDBTags = ctx.String(utils.MetricsInfluxDBTagsFlag.Name)
|
cfg.Metrics.InfluxDBTags = ctx.String(utils.MetricsInfluxDBTagsFlag.Name)
|
||||||
}
|
}
|
||||||
|
if ctx.IsSet(utils.MetricsInfluxDBIntervalFlag.Name) {
|
||||||
|
cfg.Metrics.InfluxDBInterval = ctx.Duration(utils.MetricsInfluxDBIntervalFlag.Name)
|
||||||
|
}
|
||||||
if ctx.IsSet(utils.MetricsEnableInfluxDBV2Flag.Name) {
|
if ctx.IsSet(utils.MetricsEnableInfluxDBV2Flag.Name) {
|
||||||
cfg.Metrics.EnableInfluxDBV2 = ctx.Bool(utils.MetricsEnableInfluxDBV2Flag.Name)
|
cfg.Metrics.EnableInfluxDBV2 = ctx.Bool(utils.MetricsEnableInfluxDBV2Flag.Name)
|
||||||
}
|
}
|
||||||
|
|
|
||||||
|
|
@ -30,7 +30,7 @@ import (
|
||||||
)
|
)
|
||||||
|
|
||||||
const (
|
const (
|
||||||
ipcAPIs = "admin:1.0 debug:1.0 engine:1.0 eth:1.0 miner:1.0 net:1.0 rpc:1.0 txpool:1.0 web3:1.0"
|
ipcAPIs = "admin:1.0 debug:1.0 engine:1.0 eth:1.0 miner:1.0 net:1.0 rpc:1.0 testing:1.0 txpool:1.0 web3:1.0"
|
||||||
httpAPIs = "eth:1.0 net:1.0 rpc:1.0 web3:1.0"
|
httpAPIs = "eth:1.0 net:1.0 rpc:1.0 web3:1.0"
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|
|
||||||
|
|
@ -19,6 +19,7 @@ package main
|
||||||
import (
|
import (
|
||||||
"bytes"
|
"bytes"
|
||||||
"fmt"
|
"fmt"
|
||||||
|
"math"
|
||||||
"os"
|
"os"
|
||||||
"os/signal"
|
"os/signal"
|
||||||
"path/filepath"
|
"path/filepath"
|
||||||
|
|
@ -37,6 +38,7 @@ import (
|
||||||
"github.com/ethereum/go-ethereum/core/types"
|
"github.com/ethereum/go-ethereum/core/types"
|
||||||
"github.com/ethereum/go-ethereum/crypto"
|
"github.com/ethereum/go-ethereum/crypto"
|
||||||
"github.com/ethereum/go-ethereum/ethdb"
|
"github.com/ethereum/go-ethereum/ethdb"
|
||||||
|
"github.com/ethereum/go-ethereum/internal/tablewriter"
|
||||||
"github.com/ethereum/go-ethereum/log"
|
"github.com/ethereum/go-ethereum/log"
|
||||||
"github.com/ethereum/go-ethereum/rlp"
|
"github.com/ethereum/go-ethereum/rlp"
|
||||||
"github.com/ethereum/go-ethereum/trie"
|
"github.com/ethereum/go-ethereum/trie"
|
||||||
|
|
@ -53,6 +55,23 @@ var (
|
||||||
Name: "remove.chain",
|
Name: "remove.chain",
|
||||||
Usage: "If set, selects the state data for removal",
|
Usage: "If set, selects the state data for removal",
|
||||||
}
|
}
|
||||||
|
inspectTrieTopFlag = &cli.IntFlag{
|
||||||
|
Name: "top",
|
||||||
|
Usage: "Print the top N results per ranking category",
|
||||||
|
Value: 10,
|
||||||
|
}
|
||||||
|
inspectTrieDumpPathFlag = &cli.StringFlag{
|
||||||
|
Name: "dump-path",
|
||||||
|
Usage: "Path for the trie statistics dump file",
|
||||||
|
}
|
||||||
|
inspectTrieSummarizeFlag = &cli.StringFlag{
|
||||||
|
Name: "summarize",
|
||||||
|
Usage: "Summarize an existing trie dump file (skip trie traversal)",
|
||||||
|
}
|
||||||
|
inspectTrieContractFlag = &cli.StringFlag{
|
||||||
|
Name: "contract",
|
||||||
|
Usage: "Inspect only the storage of the given contract address (skips full account trie walk)",
|
||||||
|
}
|
||||||
|
|
||||||
removedbCommand = &cli.Command{
|
removedbCommand = &cli.Command{
|
||||||
Action: removeDB,
|
Action: removeDB,
|
||||||
|
|
@ -74,6 +93,7 @@ Remove blockchain and state databases`,
|
||||||
dbCompactCmd,
|
dbCompactCmd,
|
||||||
dbGetCmd,
|
dbGetCmd,
|
||||||
dbDeleteCmd,
|
dbDeleteCmd,
|
||||||
|
dbInspectTrieCmd,
|
||||||
dbPutCmd,
|
dbPutCmd,
|
||||||
dbGetSlotsCmd,
|
dbGetSlotsCmd,
|
||||||
dbDumpFreezerIndex,
|
dbDumpFreezerIndex,
|
||||||
|
|
@ -92,6 +112,22 @@ Remove blockchain and state databases`,
|
||||||
Usage: "Inspect the storage size for each type of data in the database",
|
Usage: "Inspect the storage size for each type of data in the database",
|
||||||
Description: `This commands iterates the entire database. If the optional 'prefix' and 'start' arguments are provided, then the iteration is limited to the given subset of data.`,
|
Description: `This commands iterates the entire database. If the optional 'prefix' and 'start' arguments are provided, then the iteration is limited to the given subset of data.`,
|
||||||
}
|
}
|
||||||
|
dbInspectTrieCmd = &cli.Command{
|
||||||
|
Action: inspectTrie,
|
||||||
|
Name: "inspect-trie",
|
||||||
|
ArgsUsage: "<blocknum>",
|
||||||
|
Flags: slices.Concat([]cli.Flag{
|
||||||
|
utils.ExcludeStorageFlag,
|
||||||
|
inspectTrieTopFlag,
|
||||||
|
utils.OutputFileFlag,
|
||||||
|
inspectTrieDumpPathFlag,
|
||||||
|
inspectTrieSummarizeFlag,
|
||||||
|
inspectTrieContractFlag,
|
||||||
|
}, utils.NetworkFlags, utils.DatabaseFlags),
|
||||||
|
Usage: "Print detailed trie information about the structure of account trie and storage tries.",
|
||||||
|
Description: `This commands iterates the entrie trie-backed state. If the 'blocknum' is not specified,
|
||||||
|
the latest block number will be used by default.`,
|
||||||
|
}
|
||||||
dbCheckStateContentCmd = &cli.Command{
|
dbCheckStateContentCmd = &cli.Command{
|
||||||
Action: checkStateContent,
|
Action: checkStateContent,
|
||||||
Name: "check-state-content",
|
Name: "check-state-content",
|
||||||
|
|
@ -385,6 +421,88 @@ func checkStateContent(ctx *cli.Context) error {
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func inspectTrie(ctx *cli.Context) error {
|
||||||
|
topN := ctx.Int(inspectTrieTopFlag.Name)
|
||||||
|
if topN <= 0 {
|
||||||
|
return fmt.Errorf("invalid --%s value %d (must be > 0)", inspectTrieTopFlag.Name, topN)
|
||||||
|
}
|
||||||
|
config := &trie.InspectConfig{
|
||||||
|
NoStorage: ctx.Bool(utils.ExcludeStorageFlag.Name),
|
||||||
|
TopN: topN,
|
||||||
|
Path: ctx.String(utils.OutputFileFlag.Name),
|
||||||
|
}
|
||||||
|
|
||||||
|
if summarizePath := ctx.String(inspectTrieSummarizeFlag.Name); summarizePath != "" {
|
||||||
|
if ctx.NArg() > 0 {
|
||||||
|
return fmt.Errorf("block number argument is not supported with --%s", inspectTrieSummarizeFlag.Name)
|
||||||
|
}
|
||||||
|
config.DumpPath = summarizePath
|
||||||
|
log.Info("Summarizing trie dump", "path", summarizePath, "top", topN)
|
||||||
|
return trie.Summarize(summarizePath, config)
|
||||||
|
}
|
||||||
|
if ctx.NArg() > 1 {
|
||||||
|
return fmt.Errorf("excessive number of arguments: %v", ctx.Command.ArgsUsage)
|
||||||
|
}
|
||||||
|
|
||||||
|
stack, _ := makeConfigNode(ctx)
|
||||||
|
db := utils.MakeChainDatabase(ctx, stack, false)
|
||||||
|
defer stack.Close()
|
||||||
|
defer db.Close()
|
||||||
|
|
||||||
|
var (
|
||||||
|
trieRoot common.Hash
|
||||||
|
hash common.Hash
|
||||||
|
number uint64
|
||||||
|
)
|
||||||
|
switch {
|
||||||
|
case ctx.NArg() == 0 || ctx.Args().Get(0) == "latest":
|
||||||
|
head := rawdb.ReadHeadHeaderHash(db)
|
||||||
|
n, ok := rawdb.ReadHeaderNumber(db, head)
|
||||||
|
if !ok {
|
||||||
|
return fmt.Errorf("could not load head block hash")
|
||||||
|
}
|
||||||
|
number = n
|
||||||
|
case ctx.Args().Get(0) == "snapshot":
|
||||||
|
trieRoot = rawdb.ReadSnapshotRoot(db)
|
||||||
|
number = math.MaxUint64
|
||||||
|
default:
|
||||||
|
var err error
|
||||||
|
number, err = strconv.ParseUint(ctx.Args().Get(0), 10, 64)
|
||||||
|
if err != nil {
|
||||||
|
return fmt.Errorf("failed to parse blocknum, Args[0]: %v, err: %v", ctx.Args().Get(0), err)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
if number != math.MaxUint64 {
|
||||||
|
hash = rawdb.ReadCanonicalHash(db, number)
|
||||||
|
if hash == (common.Hash{}) {
|
||||||
|
return fmt.Errorf("canonical hash for block %d not found", number)
|
||||||
|
}
|
||||||
|
blockHeader := rawdb.ReadHeader(db, hash, number)
|
||||||
|
trieRoot = blockHeader.Root
|
||||||
|
}
|
||||||
|
if trieRoot == (common.Hash{}) {
|
||||||
|
log.Error("Empty root hash")
|
||||||
|
}
|
||||||
|
|
||||||
|
config.DumpPath = ctx.String(inspectTrieDumpPathFlag.Name)
|
||||||
|
if config.DumpPath == "" {
|
||||||
|
config.DumpPath = stack.ResolvePath("trie-dump.bin")
|
||||||
|
}
|
||||||
|
|
||||||
|
triedb := utils.MakeTrieDatabase(ctx, stack, db, false, true, false)
|
||||||
|
defer triedb.Close()
|
||||||
|
|
||||||
|
if contractAddr := ctx.String(inspectTrieContractFlag.Name); contractAddr != "" {
|
||||||
|
address := common.HexToAddress(contractAddr)
|
||||||
|
log.Info("Inspecting contract", "address", address, "root", trieRoot, "block", number)
|
||||||
|
return trie.InspectContract(triedb, db, trieRoot, address)
|
||||||
|
}
|
||||||
|
|
||||||
|
log.Info("Inspecting trie", "root", trieRoot, "block", number, "dump", config.DumpPath, "top", topN)
|
||||||
|
return trie.Inspect(triedb, trieRoot, config)
|
||||||
|
}
|
||||||
|
|
||||||
func showDBStats(db ethdb.KeyValueStater) {
|
func showDBStats(db ethdb.KeyValueStater) {
|
||||||
stats, err := db.Stat()
|
stats, err := db.Stat()
|
||||||
if err != nil {
|
if err != nil {
|
||||||
|
|
@ -759,7 +877,7 @@ func showMetaData(ctx *cli.Context) error {
|
||||||
data = append(data, []string{"headHeader.Root", fmt.Sprintf("%v", h.Root)})
|
data = append(data, []string{"headHeader.Root", fmt.Sprintf("%v", h.Root)})
|
||||||
data = append(data, []string{"headHeader.Number", fmt.Sprintf("%d (%#x)", h.Number, h.Number)})
|
data = append(data, []string{"headHeader.Number", fmt.Sprintf("%d (%#x)", h.Number, h.Number)})
|
||||||
}
|
}
|
||||||
table := rawdb.NewTableWriter(os.Stdout)
|
table := tablewriter.NewWriter(os.Stdout)
|
||||||
table.SetHeader([]string{"Field", "Value"})
|
table.SetHeader([]string{"Field", "Value"})
|
||||||
table.AppendBulk(data)
|
table.AppendBulk(data)
|
||||||
table.Render()
|
table.Render()
|
||||||
|
|
|
||||||
|
|
@ -22,7 +22,6 @@ import (
|
||||||
"os"
|
"os"
|
||||||
"slices"
|
"slices"
|
||||||
"sort"
|
"sort"
|
||||||
"strconv"
|
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
"github.com/ethereum/go-ethereum/accounts"
|
"github.com/ethereum/go-ethereum/accounts"
|
||||||
|
|
@ -216,6 +215,7 @@ var (
|
||||||
utils.MetricsInfluxDBUsernameFlag,
|
utils.MetricsInfluxDBUsernameFlag,
|
||||||
utils.MetricsInfluxDBPasswordFlag,
|
utils.MetricsInfluxDBPasswordFlag,
|
||||||
utils.MetricsInfluxDBTagsFlag,
|
utils.MetricsInfluxDBTagsFlag,
|
||||||
|
utils.MetricsInfluxDBIntervalFlag,
|
||||||
utils.MetricsEnableInfluxDBV2Flag,
|
utils.MetricsEnableInfluxDBV2Flag,
|
||||||
utils.MetricsInfluxDBTokenFlag,
|
utils.MetricsInfluxDBTokenFlag,
|
||||||
utils.MetricsInfluxDBBucketFlag,
|
utils.MetricsInfluxDBBucketFlag,
|
||||||
|
|
@ -315,18 +315,6 @@ func prepare(ctx *cli.Context) {
|
||||||
case !ctx.IsSet(utils.NetworkIdFlag.Name):
|
case !ctx.IsSet(utils.NetworkIdFlag.Name):
|
||||||
log.Info("Starting Geth on Ethereum mainnet...")
|
log.Info("Starting Geth on Ethereum mainnet...")
|
||||||
}
|
}
|
||||||
// If we're a full node on mainnet without --cache specified, bump default cache allowance
|
|
||||||
if !ctx.IsSet(utils.CacheFlag.Name) && !ctx.IsSet(utils.NetworkIdFlag.Name) {
|
|
||||||
// Make sure we're not on any supported preconfigured testnet either
|
|
||||||
if !ctx.IsSet(utils.HoleskyFlag.Name) &&
|
|
||||||
!ctx.IsSet(utils.SepoliaFlag.Name) &&
|
|
||||||
!ctx.IsSet(utils.HoodiFlag.Name) &&
|
|
||||||
!ctx.IsSet(utils.DeveloperFlag.Name) {
|
|
||||||
// Nope, we're really on mainnet. Bump that cache up!
|
|
||||||
log.Info("Bumping default cache on mainnet", "provided", ctx.Int(utils.CacheFlag.Name), "updated", 4096)
|
|
||||||
ctx.Set(utils.CacheFlag.Name, strconv.Itoa(4096))
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// geth is the main entry point into the system if no special subcommand is run.
|
// geth is the main entry point into the system if no special subcommand is run.
|
||||||
|
|
|
||||||
|
|
@ -17,7 +17,7 @@ require (
|
||||||
github.com/deckarep/golang-set/v2 v2.6.0 // indirect
|
github.com/deckarep/golang-set/v2 v2.6.0 // indirect
|
||||||
github.com/decred/dcrd/dcrec/secp256k1/v4 v4.0.1 // indirect
|
github.com/decred/dcrd/dcrec/secp256k1/v4 v4.0.1 // indirect
|
||||||
github.com/emicklei/dot v1.6.2 // indirect
|
github.com/emicklei/dot v1.6.2 // indirect
|
||||||
github.com/ethereum/c-kzg-4844/v2 v2.1.5 // indirect
|
github.com/ethereum/c-kzg-4844/v2 v2.1.6 // indirect
|
||||||
github.com/ethereum/go-bigmodexpfix v0.0.0-20250911101455-f9e208c548ab // indirect
|
github.com/ethereum/go-bigmodexpfix v0.0.0-20250911101455-f9e208c548ab // indirect
|
||||||
github.com/ferranbt/fastssz v0.1.4 // indirect
|
github.com/ferranbt/fastssz v0.1.4 // indirect
|
||||||
github.com/go-logr/logr v1.4.3 // indirect
|
github.com/go-logr/logr v1.4.3 // indirect
|
||||||
|
|
@ -31,16 +31,16 @@ require (
|
||||||
github.com/minio/sha256-simd v1.0.0 // indirect
|
github.com/minio/sha256-simd v1.0.0 // indirect
|
||||||
github.com/mitchellh/mapstructure v1.4.1 // indirect
|
github.com/mitchellh/mapstructure v1.4.1 // indirect
|
||||||
github.com/shirou/gopsutil v3.21.4-0.20210419000835-c7a38de76ee5+incompatible // indirect
|
github.com/shirou/gopsutil v3.21.4-0.20210419000835-c7a38de76ee5+incompatible // indirect
|
||||||
github.com/supranational/blst v0.3.16-0.20250831170142-f48500c1fdbe // indirect
|
github.com/supranational/blst v0.3.16 // indirect
|
||||||
github.com/tklauser/go-sysconf v0.3.12 // indirect
|
github.com/tklauser/go-sysconf v0.3.12 // indirect
|
||||||
github.com/tklauser/numcpus v0.6.1 // indirect
|
github.com/tklauser/numcpus v0.6.1 // indirect
|
||||||
go.opentelemetry.io/auto/sdk v1.2.1 // indirect
|
go.opentelemetry.io/auto/sdk v1.2.1 // indirect
|
||||||
go.opentelemetry.io/otel v1.39.0 // indirect
|
go.opentelemetry.io/otel v1.40.0 // indirect
|
||||||
go.opentelemetry.io/otel/metric v1.39.0 // indirect
|
go.opentelemetry.io/otel/metric v1.40.0 // indirect
|
||||||
go.opentelemetry.io/otel/trace v1.39.0 // indirect
|
go.opentelemetry.io/otel/trace v1.40.0 // indirect
|
||||||
golang.org/x/crypto v0.44.0 // indirect
|
golang.org/x/crypto v0.44.0 // indirect
|
||||||
golang.org/x/sync v0.18.0 // indirect
|
golang.org/x/sync v0.18.0 // indirect
|
||||||
golang.org/x/sys v0.39.0 // indirect
|
golang.org/x/sys v0.40.0 // indirect
|
||||||
gopkg.in/yaml.v2 v2.4.0 // indirect
|
gopkg.in/yaml.v2 v2.4.0 // indirect
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|
|
||||||
|
|
@ -40,8 +40,8 @@ github.com/decred/dcrd/dcrec/secp256k1/v4 v4.0.1 h1:YLtO71vCjJRCBcrPMtQ9nqBsqpA1
|
||||||
github.com/decred/dcrd/dcrec/secp256k1/v4 v4.0.1/go.mod h1:hyedUtir6IdtD/7lIxGeCxkaw7y45JueMRL4DIyJDKs=
|
github.com/decred/dcrd/dcrec/secp256k1/v4 v4.0.1/go.mod h1:hyedUtir6IdtD/7lIxGeCxkaw7y45JueMRL4DIyJDKs=
|
||||||
github.com/emicklei/dot v1.6.2 h1:08GN+DD79cy/tzN6uLCT84+2Wk9u+wvqP+Hkx/dIR8A=
|
github.com/emicklei/dot v1.6.2 h1:08GN+DD79cy/tzN6uLCT84+2Wk9u+wvqP+Hkx/dIR8A=
|
||||||
github.com/emicklei/dot v1.6.2/go.mod h1:DeV7GvQtIw4h2u73RKBkkFdvVAz0D9fzeJrgPW6gy/s=
|
github.com/emicklei/dot v1.6.2/go.mod h1:DeV7GvQtIw4h2u73RKBkkFdvVAz0D9fzeJrgPW6gy/s=
|
||||||
github.com/ethereum/c-kzg-4844/v2 v2.1.5 h1:aVtoLK5xwJ6c5RiqO8g8ptJ5KU+2Hdquf6G3aXiHh5s=
|
github.com/ethereum/c-kzg-4844/v2 v2.1.6 h1:xQymkKCT5E2Jiaoqf3v4wsNgjZLY0lRSkZn27fRjSls=
|
||||||
github.com/ethereum/c-kzg-4844/v2 v2.1.5/go.mod h1:u59hRTTah4Co6i9fDWtiCjTrblJv0UwsqZKCc0GfgUs=
|
github.com/ethereum/c-kzg-4844/v2 v2.1.6/go.mod h1:8HMkUZ5JRv4hpw/XUrYWSQNAUzhHMg2UDb/U+5m+XNw=
|
||||||
github.com/ethereum/go-bigmodexpfix v0.0.0-20250911101455-f9e208c548ab h1:rvv6MJhy07IMfEKuARQ9TKojGqLVNxQajaXEp/BoqSk=
|
github.com/ethereum/go-bigmodexpfix v0.0.0-20250911101455-f9e208c548ab h1:rvv6MJhy07IMfEKuARQ9TKojGqLVNxQajaXEp/BoqSk=
|
||||||
github.com/ethereum/go-bigmodexpfix v0.0.0-20250911101455-f9e208c548ab/go.mod h1:IuLm4IsPipXKF7CW5Lzf68PIbZ5yl7FFd74l/E0o9A8=
|
github.com/ethereum/go-bigmodexpfix v0.0.0-20250911101455-f9e208c548ab/go.mod h1:IuLm4IsPipXKF7CW5Lzf68PIbZ5yl7FFd74l/E0o9A8=
|
||||||
github.com/ferranbt/fastssz v0.1.4 h1:OCDB+dYDEQDvAgtAGnTSidK1Pe2tW3nFV40XyMkTeDY=
|
github.com/ferranbt/fastssz v0.1.4 h1:OCDB+dYDEQDvAgtAGnTSidK1Pe2tW3nFV40XyMkTeDY=
|
||||||
|
|
@ -111,22 +111,22 @@ github.com/shirou/gopsutil v3.21.4-0.20210419000835-c7a38de76ee5+incompatible h1
|
||||||
github.com/shirou/gopsutil v3.21.4-0.20210419000835-c7a38de76ee5+incompatible/go.mod h1:5b4v6he4MtMOwMlS0TUMTu2PcXUg8+E1lC7eC3UO/RA=
|
github.com/shirou/gopsutil v3.21.4-0.20210419000835-c7a38de76ee5+incompatible/go.mod h1:5b4v6he4MtMOwMlS0TUMTu2PcXUg8+E1lC7eC3UO/RA=
|
||||||
github.com/stretchr/testify v1.11.1 h1:7s2iGBzp5EwR7/aIZr8ao5+dra3wiQyKjjFuvgVKu7U=
|
github.com/stretchr/testify v1.11.1 h1:7s2iGBzp5EwR7/aIZr8ao5+dra3wiQyKjjFuvgVKu7U=
|
||||||
github.com/stretchr/testify v1.11.1/go.mod h1:wZwfW3scLgRK+23gO65QZefKpKQRnfz6sD981Nm4B6U=
|
github.com/stretchr/testify v1.11.1/go.mod h1:wZwfW3scLgRK+23gO65QZefKpKQRnfz6sD981Nm4B6U=
|
||||||
github.com/supranational/blst v0.3.16-0.20250831170142-f48500c1fdbe h1:nbdqkIGOGfUAD54q1s2YBcBz/WcsxCO9HUQ4aGV5hUw=
|
github.com/supranational/blst v0.3.16 h1:bTDadT+3fK497EvLdWRQEjiGnUtzJ7jjIUMF0jqwYhE=
|
||||||
github.com/supranational/blst v0.3.16-0.20250831170142-f48500c1fdbe/go.mod h1:jZJtfjgudtNl4en1tzwPIV3KjUnQUvG3/j+w+fVonLw=
|
github.com/supranational/blst v0.3.16/go.mod h1:jZJtfjgudtNl4en1tzwPIV3KjUnQUvG3/j+w+fVonLw=
|
||||||
github.com/tklauser/go-sysconf v0.3.12 h1:0QaGUFOdQaIVdPgfITYzaTegZvdCjmYO52cSFAEVmqU=
|
github.com/tklauser/go-sysconf v0.3.12 h1:0QaGUFOdQaIVdPgfITYzaTegZvdCjmYO52cSFAEVmqU=
|
||||||
github.com/tklauser/go-sysconf v0.3.12/go.mod h1:Ho14jnntGE1fpdOqQEEaiKRpvIavV0hSfmBq8nJbHYI=
|
github.com/tklauser/go-sysconf v0.3.12/go.mod h1:Ho14jnntGE1fpdOqQEEaiKRpvIavV0hSfmBq8nJbHYI=
|
||||||
github.com/tklauser/numcpus v0.6.1 h1:ng9scYS7az0Bk4OZLvrNXNSAO2Pxr1XXRAPyjhIx+Fk=
|
github.com/tklauser/numcpus v0.6.1 h1:ng9scYS7az0Bk4OZLvrNXNSAO2Pxr1XXRAPyjhIx+Fk=
|
||||||
github.com/tklauser/numcpus v0.6.1/go.mod h1:1XfjsgE2zo8GVw7POkMbHENHzVg3GzmoZ9fESEdAacY=
|
github.com/tklauser/numcpus v0.6.1/go.mod h1:1XfjsgE2zo8GVw7POkMbHENHzVg3GzmoZ9fESEdAacY=
|
||||||
go.opentelemetry.io/auto/sdk v1.2.1 h1:jXsnJ4Lmnqd11kwkBV2LgLoFMZKizbCi5fNZ/ipaZ64=
|
go.opentelemetry.io/auto/sdk v1.2.1 h1:jXsnJ4Lmnqd11kwkBV2LgLoFMZKizbCi5fNZ/ipaZ64=
|
||||||
go.opentelemetry.io/auto/sdk v1.2.1/go.mod h1:KRTj+aOaElaLi+wW1kO/DZRXwkF4C5xPbEe3ZiIhN7Y=
|
go.opentelemetry.io/auto/sdk v1.2.1/go.mod h1:KRTj+aOaElaLi+wW1kO/DZRXwkF4C5xPbEe3ZiIhN7Y=
|
||||||
go.opentelemetry.io/otel v1.39.0 h1:8yPrr/S0ND9QEfTfdP9V+SiwT4E0G7Y5MO7p85nis48=
|
go.opentelemetry.io/otel v1.40.0 h1:oA5YeOcpRTXq6NN7frwmwFR0Cn3RhTVZvXsP4duvCms=
|
||||||
go.opentelemetry.io/otel v1.39.0/go.mod h1:kLlFTywNWrFyEdH0oj2xK0bFYZtHRYUdv1NklR/tgc8=
|
go.opentelemetry.io/otel v1.40.0/go.mod h1:IMb+uXZUKkMXdPddhwAHm6UfOwJyh4ct1ybIlV14J0g=
|
||||||
go.opentelemetry.io/otel/metric v1.39.0 h1:d1UzonvEZriVfpNKEVmHXbdf909uGTOQjA0HF0Ls5Q0=
|
go.opentelemetry.io/otel/metric v1.40.0 h1:rcZe317KPftE2rstWIBitCdVp89A2HqjkxR3c11+p9g=
|
||||||
go.opentelemetry.io/otel/metric v1.39.0/go.mod h1:jrZSWL33sD7bBxg1xjrqyDjnuzTUB0x1nBERXd7Ftcs=
|
go.opentelemetry.io/otel/metric v1.40.0/go.mod h1:ib/crwQH7N3r5kfiBZQbwrTge743UDc7DTFVZrrXnqc=
|
||||||
go.opentelemetry.io/otel/sdk v1.39.0 h1:nMLYcjVsvdui1B/4FRkwjzoRVsMK8uL/cj0OyhKzt18=
|
go.opentelemetry.io/otel/sdk v1.40.0 h1:KHW/jUzgo6wsPh9At46+h4upjtccTmuZCFAc9OJ71f8=
|
||||||
go.opentelemetry.io/otel/sdk v1.39.0/go.mod h1:vDojkC4/jsTJsE+kh+LXYQlbL8CgrEcwmt1ENZszdJE=
|
go.opentelemetry.io/otel/sdk v1.40.0/go.mod h1:Ph7EFdYvxq72Y8Li9q8KebuYUr2KoeyHx0DRMKrYBUE=
|
||||||
go.opentelemetry.io/otel/trace v1.39.0 h1:2d2vfpEDmCJ5zVYz7ijaJdOF59xLomrvj7bjt6/qCJI=
|
go.opentelemetry.io/otel/trace v1.40.0 h1:WA4etStDttCSYuhwvEa8OP8I5EWu24lkOzp+ZYblVjw=
|
||||||
go.opentelemetry.io/otel/trace v1.39.0/go.mod h1:88w4/PnZSazkGzz/w84VHpQafiU4EtqqlVdxWy+rNOA=
|
go.opentelemetry.io/otel/trace v1.40.0/go.mod h1:zeAhriXecNGP/s2SEG3+Y8X9ujcJOTqQ5RgdEJcawiA=
|
||||||
golang.org/x/crypto v0.44.0 h1:A97SsFvM3AIwEEmTBiaxPPTYpDC47w720rdiiUvgoAU=
|
golang.org/x/crypto v0.44.0 h1:A97SsFvM3AIwEEmTBiaxPPTYpDC47w720rdiiUvgoAU=
|
||||||
golang.org/x/crypto v0.44.0/go.mod h1:013i+Nw79BMiQiMsOPcVCB5ZIJbYkerPrGnOa00tvmc=
|
golang.org/x/crypto v0.44.0/go.mod h1:013i+Nw79BMiQiMsOPcVCB5ZIJbYkerPrGnOa00tvmc=
|
||||||
golang.org/x/exp v0.0.0-20230626212559-97b1e661b5df h1:UA2aFVmmsIlefxMk29Dp2juaUSth8Pyn3Tq5Y5mJGME=
|
golang.org/x/exp v0.0.0-20230626212559-97b1e661b5df h1:UA2aFVmmsIlefxMk29Dp2juaUSth8Pyn3Tq5Y5mJGME=
|
||||||
|
|
@ -137,8 +137,8 @@ golang.org/x/sys v0.0.0-20190916202348-b4ddaad3f8a3/go.mod h1:h1NjWce9XRLGQEsW7w
|
||||||
golang.org/x/sys v0.1.0/go.mod h1:oPkhp1MJrh7nUepCBck5+mAzfO9JrbApNNgaTdGDITg=
|
golang.org/x/sys v0.1.0/go.mod h1:oPkhp1MJrh7nUepCBck5+mAzfO9JrbApNNgaTdGDITg=
|
||||||
golang.org/x/sys v0.8.0/go.mod h1:oPkhp1MJrh7nUepCBck5+mAzfO9JrbApNNgaTdGDITg=
|
golang.org/x/sys v0.8.0/go.mod h1:oPkhp1MJrh7nUepCBck5+mAzfO9JrbApNNgaTdGDITg=
|
||||||
golang.org/x/sys v0.11.0/go.mod h1:oPkhp1MJrh7nUepCBck5+mAzfO9JrbApNNgaTdGDITg=
|
golang.org/x/sys v0.11.0/go.mod h1:oPkhp1MJrh7nUepCBck5+mAzfO9JrbApNNgaTdGDITg=
|
||||||
golang.org/x/sys v0.39.0 h1:CvCKL8MeisomCi6qNZ+wbb0DN9E5AATixKsvNtMoMFk=
|
golang.org/x/sys v0.40.0 h1:DBZZqJ2Rkml6QMQsZywtnjnnGvHza6BTfYFWY9kjEWQ=
|
||||||
golang.org/x/sys v0.39.0/go.mod h1:OgkHotnGiDImocRcuBABYBEXf8A9a87e/uXjp9XT3ks=
|
golang.org/x/sys v0.40.0/go.mod h1:OgkHotnGiDImocRcuBABYBEXf8A9a87e/uXjp9XT3ks=
|
||||||
golang.org/x/text v0.31.0 h1:aC8ghyu4JhP8VojJ2lEHBnochRno1sgL6nEi9WGFGMM=
|
golang.org/x/text v0.31.0 h1:aC8ghyu4JhP8VojJ2lEHBnochRno1sgL6nEi9WGFGMM=
|
||||||
golang.org/x/text v0.31.0/go.mod h1:tKRAlv61yKIjGGHX/4tP1LTbc13YSec1pxVEWXzfoeM=
|
golang.org/x/text v0.31.0/go.mod h1:tKRAlv61yKIjGGHX/4tP1LTbc13YSec1pxVEWXzfoeM=
|
||||||
google.golang.org/protobuf v1.36.11 h1:fV6ZwhNocDyBLK0dj+fg8ektcVegBBuEolpbTQyBNVE=
|
google.golang.org/protobuf v1.36.11 h1:fV6ZwhNocDyBLK0dj+fg8ektcVegBBuEolpbTQyBNVE=
|
||||||
|
|
|
||||||
|
|
@ -218,6 +218,16 @@ var (
|
||||||
Usage: "Max number of elements (0 = no limit)",
|
Usage: "Max number of elements (0 = no limit)",
|
||||||
Value: 0,
|
Value: 0,
|
||||||
}
|
}
|
||||||
|
TopFlag = &cli.IntFlag{
|
||||||
|
Name: "top",
|
||||||
|
Usage: "Print the top N results",
|
||||||
|
Value: 5,
|
||||||
|
}
|
||||||
|
OutputFileFlag = &cli.StringFlag{
|
||||||
|
Name: "output",
|
||||||
|
Usage: "Writes the result in json to the output",
|
||||||
|
Value: "",
|
||||||
|
}
|
||||||
|
|
||||||
SnapshotFlag = &cli.BoolFlag{
|
SnapshotFlag = &cli.BoolFlag{
|
||||||
Name: "snapshot",
|
Name: "snapshot",
|
||||||
|
|
@ -481,7 +491,7 @@ var (
|
||||||
CacheFlag = &cli.IntFlag{
|
CacheFlag = &cli.IntFlag{
|
||||||
Name: "cache",
|
Name: "cache",
|
||||||
Usage: "Megabytes of memory allocated to internal caching (default = 4096 mainnet full node, 128 light mode)",
|
Usage: "Megabytes of memory allocated to internal caching (default = 4096 mainnet full node, 128 light mode)",
|
||||||
Value: 1024,
|
Value: 4096,
|
||||||
Category: flags.PerfCategory,
|
Category: flags.PerfCategory,
|
||||||
}
|
}
|
||||||
CacheDatabaseFlag = &cli.IntFlag{
|
CacheDatabaseFlag = &cli.IntFlag{
|
||||||
|
|
@ -1016,6 +1026,13 @@ Please note that --` + MetricsHTTPFlag.Name + ` must be set to start the server.
|
||||||
Category: flags.MetricsCategory,
|
Category: flags.MetricsCategory,
|
||||||
}
|
}
|
||||||
|
|
||||||
|
MetricsInfluxDBIntervalFlag = &cli.DurationFlag{
|
||||||
|
Name: "metrics.influxdb.interval",
|
||||||
|
Usage: "Interval between metrics reports to InfluxDB (with time unit, e.g. 10s)",
|
||||||
|
Value: metrics.DefaultConfig.InfluxDBInterval,
|
||||||
|
Category: flags.MetricsCategory,
|
||||||
|
}
|
||||||
|
|
||||||
MetricsEnableInfluxDBV2Flag = &cli.BoolFlag{
|
MetricsEnableInfluxDBV2Flag = &cli.BoolFlag{
|
||||||
Name: "metrics.influxdbv2",
|
Name: "metrics.influxdbv2",
|
||||||
Usage: "Enable metrics export/push to an external InfluxDB v2 database",
|
Usage: "Enable metrics export/push to an external InfluxDB v2 database",
|
||||||
|
|
@ -2246,13 +2263,14 @@ func SetupMetrics(cfg *metrics.Config) {
|
||||||
bucket = cfg.InfluxDBBucket
|
bucket = cfg.InfluxDBBucket
|
||||||
organization = cfg.InfluxDBOrganization
|
organization = cfg.InfluxDBOrganization
|
||||||
tagsMap = SplitTagsFlag(cfg.InfluxDBTags)
|
tagsMap = SplitTagsFlag(cfg.InfluxDBTags)
|
||||||
|
interval = cfg.InfluxDBInterval
|
||||||
)
|
)
|
||||||
if enableExport {
|
if enableExport {
|
||||||
log.Info("Enabling metrics export to InfluxDB")
|
log.Info("Enabling metrics export to InfluxDB", "interval", interval)
|
||||||
go influxdb.InfluxDBWithTags(metrics.DefaultRegistry, 10*time.Second, endpoint, database, username, password, "geth.", tagsMap)
|
go influxdb.InfluxDBWithTags(metrics.DefaultRegistry, interval, endpoint, database, username, password, "geth.", tagsMap)
|
||||||
} else if enableExportV2 {
|
} else if enableExportV2 {
|
||||||
log.Info("Enabling metrics export to InfluxDB (v2)")
|
log.Info("Enabling metrics export to InfluxDB (v2)", "interval", interval)
|
||||||
go influxdb.InfluxDBV2WithTags(metrics.DefaultRegistry, 10*time.Second, endpoint, token, bucket, organization, "geth.", tagsMap)
|
go influxdb.InfluxDBV2WithTags(metrics.DefaultRegistry, interval, endpoint, token, bucket, organization, "geth.", tagsMap)
|
||||||
}
|
}
|
||||||
|
|
||||||
// Expvar exporter.
|
// Expvar exporter.
|
||||||
|
|
@ -2457,8 +2475,6 @@ func MakeChain(ctx *cli.Context, stack *node.Node, readonly bool) (*core.BlockCh
|
||||||
}
|
}
|
||||||
vmcfg := vm.Config{
|
vmcfg := vm.Config{
|
||||||
EnablePreimageRecording: ctx.Bool(VMEnableDebugFlag.Name),
|
EnablePreimageRecording: ctx.Bool(VMEnableDebugFlag.Name),
|
||||||
EnableWitnessStats: ctx.Bool(VMWitnessStatsFlag.Name),
|
|
||||||
StatelessSelfValidation: ctx.Bool(VMStatelessSelfValidationFlag.Name) || ctx.Bool(VMWitnessStatsFlag.Name),
|
|
||||||
}
|
}
|
||||||
if ctx.IsSet(VMTraceFlag.Name) {
|
if ctx.IsSet(VMTraceFlag.Name) {
|
||||||
if name := ctx.String(VMTraceFlag.Name); name != "" {
|
if name := ctx.String(VMTraceFlag.Name); name != "" {
|
||||||
|
|
@ -2472,6 +2488,9 @@ func MakeChain(ctx *cli.Context, stack *node.Node, readonly bool) (*core.BlockCh
|
||||||
}
|
}
|
||||||
options.VmConfig = vmcfg
|
options.VmConfig = vmcfg
|
||||||
|
|
||||||
|
options.StatelessSelfValidation = ctx.Bool(VMStatelessSelfValidationFlag.Name) || ctx.Bool(VMWitnessStatsFlag.Name)
|
||||||
|
options.EnableWitnessStats = ctx.Bool(VMWitnessStatsFlag.Name)
|
||||||
|
|
||||||
chain, err := core.NewBlockChain(chainDb, gspec, engine, options)
|
chain, err := core.NewBlockChain(chainDb, gspec, engine, options)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
Fatalf("Can't create BlockChain: %v", err)
|
Fatalf("Can't create BlockChain: %v", err)
|
||||||
|
|
|
||||||
|
|
@ -272,6 +272,14 @@ func (beacon *Beacon) verifyHeader(chain consensus.ChainHeaderReader, header, pa
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
amsterdam := chain.Config().IsAmsterdam(header.Number, header.Time)
|
||||||
|
if amsterdam && header.SlotNumber == nil {
|
||||||
|
return errors.New("header is missing slotNumber")
|
||||||
|
}
|
||||||
|
if !amsterdam && header.SlotNumber != nil {
|
||||||
|
return fmt.Errorf("invalid slotNumber: have %d, expected nil", *header.SlotNumber)
|
||||||
|
}
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
||||||
|
|
@ -310,6 +310,8 @@ func (c *Clique) verifyHeader(chain consensus.ChainHeaderReader, header *types.H
|
||||||
return fmt.Errorf("invalid blobGasUsed: have %d, expected nil", *header.BlobGasUsed)
|
return fmt.Errorf("invalid blobGasUsed: have %d, expected nil", *header.BlobGasUsed)
|
||||||
case header.ParentBeaconRoot != nil:
|
case header.ParentBeaconRoot != nil:
|
||||||
return fmt.Errorf("invalid parentBeaconRoot, have %#x, expected nil", *header.ParentBeaconRoot)
|
return fmt.Errorf("invalid parentBeaconRoot, have %#x, expected nil", *header.ParentBeaconRoot)
|
||||||
|
case header.SlotNumber != nil:
|
||||||
|
return fmt.Errorf("invalid slotNumber, have %#x, expected nil", *header.SlotNumber)
|
||||||
}
|
}
|
||||||
// All basic checks passed, verify cascading fields
|
// All basic checks passed, verify cascading fields
|
||||||
return c.verifyCascadingFields(chain, header, parents)
|
return c.verifyCascadingFields(chain, header, parents)
|
||||||
|
|
@ -694,6 +696,9 @@ func encodeSigHeader(w io.Writer, header *types.Header) {
|
||||||
if header.ParentBeaconRoot != nil {
|
if header.ParentBeaconRoot != nil {
|
||||||
panic("unexpected parent beacon root value in clique")
|
panic("unexpected parent beacon root value in clique")
|
||||||
}
|
}
|
||||||
|
if header.SlotNumber != nil {
|
||||||
|
panic("unexpected slot number value in clique")
|
||||||
|
}
|
||||||
if err := rlp.Encode(w, enc); err != nil {
|
if err := rlp.Encode(w, enc); err != nil {
|
||||||
panic("can't encode: " + err.Error())
|
panic("can't encode: " + err.Error())
|
||||||
}
|
}
|
||||||
|
|
|
||||||
|
|
@ -283,6 +283,8 @@ func (ethash *Ethash) verifyHeader(chain consensus.ChainHeaderReader, header, pa
|
||||||
return fmt.Errorf("invalid blobGasUsed: have %d, expected nil", *header.BlobGasUsed)
|
return fmt.Errorf("invalid blobGasUsed: have %d, expected nil", *header.BlobGasUsed)
|
||||||
case header.ParentBeaconRoot != nil:
|
case header.ParentBeaconRoot != nil:
|
||||||
return fmt.Errorf("invalid parentBeaconRoot, have %#x, expected nil", *header.ParentBeaconRoot)
|
return fmt.Errorf("invalid parentBeaconRoot, have %#x, expected nil", *header.ParentBeaconRoot)
|
||||||
|
case header.SlotNumber != nil:
|
||||||
|
return fmt.Errorf("invalid slotNumber, have %#x, expected nil", *header.SlotNumber)
|
||||||
}
|
}
|
||||||
// Add some fake checks for tests
|
// Add some fake checks for tests
|
||||||
if ethash.fakeDelay != nil {
|
if ethash.fakeDelay != nil {
|
||||||
|
|
@ -559,6 +561,9 @@ func (ethash *Ethash) SealHash(header *types.Header) (hash common.Hash) {
|
||||||
if header.ParentBeaconRoot != nil {
|
if header.ParentBeaconRoot != nil {
|
||||||
panic("parent beacon root set on ethash")
|
panic("parent beacon root set on ethash")
|
||||||
}
|
}
|
||||||
|
if header.SlotNumber != nil {
|
||||||
|
panic("slot number set on ethash")
|
||||||
|
}
|
||||||
rlp.Encode(hasher, enc)
|
rlp.Encode(hasher, enc)
|
||||||
hasher.Sum(hash[:0])
|
hasher.Sum(hash[:0])
|
||||||
return hash
|
return hash
|
||||||
|
|
|
||||||
|
|
@ -219,6 +219,10 @@ type BlockChainConfig struct {
|
||||||
// detailed statistics will be logged. Negative value means disabled (default),
|
// detailed statistics will be logged. Negative value means disabled (default),
|
||||||
// zero logs all blocks, positive value filters blocks by execution time.
|
// zero logs all blocks, positive value filters blocks by execution time.
|
||||||
SlowBlockThreshold time.Duration
|
SlowBlockThreshold time.Duration
|
||||||
|
|
||||||
|
// Execution configs
|
||||||
|
StatelessSelfValidation bool // Generate execution witnesses and self-check against them (testing purpose)
|
||||||
|
EnableWitnessStats bool // Whether trie access statistics collection is enabled
|
||||||
}
|
}
|
||||||
|
|
||||||
// DefaultConfig returns the default config.
|
// DefaultConfig returns the default config.
|
||||||
|
|
@ -1279,6 +1283,8 @@ func (bc *BlockChain) ExportN(w io.Writer, first uint64, last uint64) error {
|
||||||
func (bc *BlockChain) writeHeadBlock(block *types.Block) {
|
func (bc *BlockChain) writeHeadBlock(block *types.Block) {
|
||||||
// Add the block to the canonical chain number scheme and mark as the head
|
// Add the block to the canonical chain number scheme and mark as the head
|
||||||
batch := bc.db.NewBatch()
|
batch := bc.db.NewBatch()
|
||||||
|
defer batch.Close()
|
||||||
|
|
||||||
rawdb.WriteHeadHeaderHash(batch, block.Hash())
|
rawdb.WriteHeadHeaderHash(batch, block.Hash())
|
||||||
rawdb.WriteHeadFastBlockHash(batch, block.Hash())
|
rawdb.WriteHeadFastBlockHash(batch, block.Hash())
|
||||||
rawdb.WriteCanonicalHash(batch, block.Hash(), block.NumberU64())
|
rawdb.WriteCanonicalHash(batch, block.Hash(), block.NumberU64())
|
||||||
|
|
@ -1653,6 +1659,8 @@ func (bc *BlockChain) writeBlockWithState(block *types.Block, receipts []*types.
|
||||||
batch = bc.db.NewBatch()
|
batch = bc.db.NewBatch()
|
||||||
start = time.Now()
|
start = time.Now()
|
||||||
)
|
)
|
||||||
|
defer batch.Close()
|
||||||
|
|
||||||
rawdb.WriteBlock(batch, block)
|
rawdb.WriteBlock(batch, block)
|
||||||
rawdb.WriteReceipts(batch, block.Hash(), block.NumberU64(), receipts)
|
rawdb.WriteReceipts(batch, block.Hash(), block.NumberU64(), receipts)
|
||||||
rawdb.WritePreimages(batch, statedb.Preimages())
|
rawdb.WritePreimages(batch, statedb.Preimages())
|
||||||
|
|
@ -1990,7 +1998,15 @@ func (bc *BlockChain) insertChain(ctx context.Context, chain types.Blocks, setHe
|
||||||
}
|
}
|
||||||
// The traced section of block import.
|
// The traced section of block import.
|
||||||
start := time.Now()
|
start := time.Now()
|
||||||
res, err := bc.ProcessBlock(ctx, parent.Root, block, setHead, makeWitness && len(chain) == 1)
|
config := ExecuteConfig{
|
||||||
|
WriteState: true,
|
||||||
|
WriteHead: setHead,
|
||||||
|
EnableTracer: true,
|
||||||
|
MakeWitness: makeWitness && len(chain) == 1,
|
||||||
|
StatelessSelfValidation: bc.cfg.StatelessSelfValidation,
|
||||||
|
EnableWitnessStats: bc.cfg.EnableWitnessStats,
|
||||||
|
}
|
||||||
|
res, err := bc.ProcessBlock(ctx, parent.Root, block, config)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, it.index, err
|
return nil, it.index, err
|
||||||
}
|
}
|
||||||
|
|
@ -2073,9 +2089,36 @@ func (bpr *blockProcessingResult) Stats() *ExecuteStats {
|
||||||
return bpr.stats
|
return bpr.stats
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// ExecuteConfig defines optional behaviors during execution.
|
||||||
|
type ExecuteConfig struct {
|
||||||
|
// WriteState controls whether the computed state changes are persisted to
|
||||||
|
// the underlying storage. If false, execution is performed in-memory only.
|
||||||
|
WriteState bool
|
||||||
|
|
||||||
|
// WriteHead indicates whether the execution result should update the canonical
|
||||||
|
// chain head. It's only relevant with WriteState == True.
|
||||||
|
WriteHead bool
|
||||||
|
|
||||||
|
// EnableTracer enables execution tracing. This is typically used for debugging
|
||||||
|
// or analysis and may significantly impact performance.
|
||||||
|
EnableTracer bool
|
||||||
|
|
||||||
|
// MakeWitness indicates whether to generate execution witness data during
|
||||||
|
// execution. Enabling this may introduce additional memory and CPU overhead.
|
||||||
|
MakeWitness bool
|
||||||
|
|
||||||
|
// StatelessSelfValidation indicates whether the execution witnesses generation
|
||||||
|
// and self-validation (testing purpose) is enabled.
|
||||||
|
StatelessSelfValidation bool
|
||||||
|
|
||||||
|
// EnableWitnessStats indicates whether to enable collection of witness trie
|
||||||
|
// access statistics
|
||||||
|
EnableWitnessStats bool
|
||||||
|
}
|
||||||
|
|
||||||
// ProcessBlock executes and validates the given block. If there was no error
|
// ProcessBlock executes and validates the given block. If there was no error
|
||||||
// it writes the block and associated state to database.
|
// it writes the block and associated state to database.
|
||||||
func (bc *BlockChain) ProcessBlock(ctx context.Context, parentRoot common.Hash, block *types.Block, setHead bool, makeWitness bool) (result *blockProcessingResult, blockEndErr error) {
|
func (bc *BlockChain) ProcessBlock(ctx context.Context, parentRoot common.Hash, block *types.Block, config ExecuteConfig) (result *blockProcessingResult, blockEndErr error) {
|
||||||
var (
|
var (
|
||||||
err error
|
err error
|
||||||
startTime = time.Now()
|
startTime = time.Now()
|
||||||
|
|
@ -2138,12 +2181,12 @@ func (bc *BlockChain) ProcessBlock(ctx context.Context, parentRoot common.Hash,
|
||||||
// Generate witnesses either if we're self-testing, or if it's the
|
// Generate witnesses either if we're self-testing, or if it's the
|
||||||
// only block being inserted. A bit crude, but witnesses are huge,
|
// only block being inserted. A bit crude, but witnesses are huge,
|
||||||
// so we refuse to make an entire chain of them.
|
// so we refuse to make an entire chain of them.
|
||||||
if bc.cfg.VmConfig.StatelessSelfValidation || makeWitness {
|
if config.StatelessSelfValidation || config.MakeWitness {
|
||||||
witness, err = stateless.NewWitness(block.Header(), bc)
|
witness, err = stateless.NewWitness(block.Header(), bc)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
if bc.cfg.VmConfig.EnableWitnessStats {
|
if config.EnableWitnessStats {
|
||||||
witnessStats = stateless.NewWitnessStats()
|
witnessStats = stateless.NewWitnessStats()
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
@ -2151,17 +2194,20 @@ func (bc *BlockChain) ProcessBlock(ctx context.Context, parentRoot common.Hash,
|
||||||
defer statedb.StopPrefetcher()
|
defer statedb.StopPrefetcher()
|
||||||
}
|
}
|
||||||
|
|
||||||
if bc.logger != nil && bc.logger.OnBlockStart != nil {
|
// Instrument the blockchain tracing
|
||||||
bc.logger.OnBlockStart(tracing.BlockEvent{
|
if config.EnableTracer {
|
||||||
Block: block,
|
if bc.logger != nil && bc.logger.OnBlockStart != nil {
|
||||||
Finalized: bc.CurrentFinalBlock(),
|
bc.logger.OnBlockStart(tracing.BlockEvent{
|
||||||
Safe: bc.CurrentSafeBlock(),
|
Block: block,
|
||||||
})
|
Finalized: bc.CurrentFinalBlock(),
|
||||||
}
|
Safe: bc.CurrentSafeBlock(),
|
||||||
if bc.logger != nil && bc.logger.OnBlockEnd != nil {
|
})
|
||||||
defer func() {
|
}
|
||||||
bc.logger.OnBlockEnd(blockEndErr)
|
if bc.logger != nil && bc.logger.OnBlockEnd != nil {
|
||||||
}()
|
defer func() {
|
||||||
|
bc.logger.OnBlockEnd(blockEndErr)
|
||||||
|
}()
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Process block using the parent state as reference point
|
// Process block using the parent state as reference point
|
||||||
|
|
@ -2191,7 +2237,7 @@ func (bc *BlockChain) ProcessBlock(ctx context.Context, parentRoot common.Hash,
|
||||||
// witness builder/runner, which would otherwise be impossible due to the
|
// witness builder/runner, which would otherwise be impossible due to the
|
||||||
// various invalid chain states/behaviors being contained in those tests.
|
// various invalid chain states/behaviors being contained in those tests.
|
||||||
xvstart := time.Now()
|
xvstart := time.Now()
|
||||||
if witness := statedb.Witness(); witness != nil && bc.cfg.VmConfig.StatelessSelfValidation {
|
if witness := statedb.Witness(); witness != nil && config.StatelessSelfValidation {
|
||||||
log.Warn("Running stateless self-validation", "block", block.Number(), "hash", block.Hash())
|
log.Warn("Running stateless self-validation", "block", block.Number(), "hash", block.Hash())
|
||||||
|
|
||||||
// Remove critical computed fields from the block to force true recalculation
|
// Remove critical computed fields from the block to force true recalculation
|
||||||
|
|
@ -2244,31 +2290,29 @@ func (bc *BlockChain) ProcessBlock(ctx context.Context, parentRoot common.Hash,
|
||||||
stats.CrossValidation = xvtime // The time spent on stateless cross validation
|
stats.CrossValidation = xvtime // The time spent on stateless cross validation
|
||||||
|
|
||||||
// Write the block to the chain and get the status.
|
// Write the block to the chain and get the status.
|
||||||
var (
|
var status WriteStatus
|
||||||
wstart = time.Now()
|
if config.WriteState {
|
||||||
status WriteStatus
|
wstart := time.Now()
|
||||||
)
|
if !config.WriteHead {
|
||||||
if !setHead {
|
// Don't set the head, only insert the block
|
||||||
// Don't set the head, only insert the block
|
err = bc.writeBlockWithState(block, res.Receipts, statedb)
|
||||||
err = bc.writeBlockWithState(block, res.Receipts, statedb)
|
} else {
|
||||||
} else {
|
status, err = bc.writeBlockAndSetHead(block, res.Receipts, res.Logs, statedb, false)
|
||||||
status, err = bc.writeBlockAndSetHead(block, res.Receipts, res.Logs, statedb, false)
|
}
|
||||||
}
|
if err != nil {
|
||||||
if err != nil {
|
return nil, err
|
||||||
return nil, err
|
}
|
||||||
|
// Update the metrics touched during block commit
|
||||||
|
stats.AccountCommits = statedb.AccountCommits // Account commits are complete, we can mark them
|
||||||
|
stats.StorageCommits = statedb.StorageCommits // Storage commits are complete, we can mark them
|
||||||
|
stats.SnapshotCommit = statedb.SnapshotCommits // Snapshot commits are complete, we can mark them
|
||||||
|
stats.TrieDBCommit = statedb.TrieDBCommits // Trie database commits are complete, we can mark them
|
||||||
|
stats.BlockWrite = time.Since(wstart) - max(statedb.AccountCommits, statedb.StorageCommits) /* concurrent */ - statedb.SnapshotCommits - statedb.TrieDBCommits
|
||||||
}
|
}
|
||||||
// Report the collected witness statistics
|
// Report the collected witness statistics
|
||||||
if witnessStats != nil {
|
if witnessStats != nil {
|
||||||
witnessStats.ReportMetrics(block.NumberU64())
|
witnessStats.ReportMetrics(block.NumberU64())
|
||||||
}
|
}
|
||||||
|
|
||||||
// Update the metrics touched during block commit
|
|
||||||
stats.AccountCommits = statedb.AccountCommits // Account commits are complete, we can mark them
|
|
||||||
stats.StorageCommits = statedb.StorageCommits // Storage commits are complete, we can mark them
|
|
||||||
stats.SnapshotCommit = statedb.SnapshotCommits // Snapshot commits are complete, we can mark them
|
|
||||||
stats.TrieDBCommit = statedb.TrieDBCommits // Trie database commits are complete, we can mark them
|
|
||||||
stats.BlockWrite = time.Since(wstart) - max(statedb.AccountCommits, statedb.StorageCommits) /* concurrent */ - statedb.SnapshotCommits - statedb.TrieDBCommits
|
|
||||||
|
|
||||||
elapsed := time.Since(startTime) + 1 // prevent zero division
|
elapsed := time.Since(startTime) + 1 // prevent zero division
|
||||||
stats.TotalTime = elapsed
|
stats.TotalTime = elapsed
|
||||||
stats.MgasPerSecond = float64(res.GasUsed) * 1000 / float64(elapsed)
|
stats.MgasPerSecond = float64(res.GasUsed) * 1000 / float64(elapsed)
|
||||||
|
|
@ -2626,6 +2670,8 @@ func (bc *BlockChain) reorg(oldHead *types.Header, newHead *types.Header) error
|
||||||
// Delete useless indexes right now which includes the non-canonical
|
// Delete useless indexes right now which includes the non-canonical
|
||||||
// transaction indexes, canonical chain indexes which above the head.
|
// transaction indexes, canonical chain indexes which above the head.
|
||||||
batch := bc.db.NewBatch()
|
batch := bc.db.NewBatch()
|
||||||
|
defer batch.Close()
|
||||||
|
|
||||||
for _, tx := range types.HashDifference(deletedTxs, rebirthTxs) {
|
for _, tx := range types.HashDifference(deletedTxs, rebirthTxs) {
|
||||||
rawdb.DeleteTxLookupEntry(batch, tx)
|
rawdb.DeleteTxLookupEntry(batch, tx)
|
||||||
}
|
}
|
||||||
|
|
|
||||||
|
|
@ -63,7 +63,7 @@ func (b *BlockGen) SetCoinbase(addr common.Address) {
|
||||||
panic("coinbase can only be set once")
|
panic("coinbase can only be set once")
|
||||||
}
|
}
|
||||||
b.header.Coinbase = addr
|
b.header.Coinbase = addr
|
||||||
b.gasPool = new(GasPool).AddGas(b.header.GasLimit)
|
b.gasPool = NewGasPool(b.header.GasLimit)
|
||||||
}
|
}
|
||||||
|
|
||||||
// SetExtra sets the extra data field of the generated block.
|
// SetExtra sets the extra data field of the generated block.
|
||||||
|
|
@ -117,10 +117,12 @@ func (b *BlockGen) addTx(bc *BlockChain, vmConfig vm.Config, tx *types.Transacti
|
||||||
evm = vm.NewEVM(blockContext, b.statedb, b.cm.config, vmConfig)
|
evm = vm.NewEVM(blockContext, b.statedb, b.cm.config, vmConfig)
|
||||||
)
|
)
|
||||||
b.statedb.SetTxContext(tx.Hash(), len(b.txs))
|
b.statedb.SetTxContext(tx.Hash(), len(b.txs))
|
||||||
receipt, err := ApplyTransaction(evm, b.gasPool, b.statedb, b.header, tx, &b.header.GasUsed)
|
receipt, err := ApplyTransaction(evm, b.gasPool, b.statedb, b.header, tx)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
panic(err)
|
panic(err)
|
||||||
}
|
}
|
||||||
|
b.header.GasUsed = b.gasPool.Used()
|
||||||
|
|
||||||
// Merge the tx-local access event into the "block-local" one, in order to collect
|
// Merge the tx-local access event into the "block-local" one, in order to collect
|
||||||
// all values, so that the witness can be built.
|
// all values, so that the witness can be built.
|
||||||
if b.statedb.Database().TrieDB().IsVerkle() {
|
if b.statedb.Database().TrieDB().IsVerkle() {
|
||||||
|
|
|
||||||
|
|
@ -58,6 +58,10 @@ var (
|
||||||
// by a transaction is higher than what's left in the block.
|
// by a transaction is higher than what's left in the block.
|
||||||
ErrGasLimitReached = errors.New("gas limit reached")
|
ErrGasLimitReached = errors.New("gas limit reached")
|
||||||
|
|
||||||
|
// ErrGasLimitOverflow is returned by the gas pool if the remaining gas
|
||||||
|
// exceeds the maximum value of uint64.
|
||||||
|
ErrGasLimitOverflow = errors.New("gas limit overflow")
|
||||||
|
|
||||||
// ErrInsufficientFundsForTransfer is returned if the transaction sender doesn't
|
// ErrInsufficientFundsForTransfer is returned if the transaction sender doesn't
|
||||||
// have enough funds for transfer(topmost call only).
|
// have enough funds for transfer(topmost call only).
|
||||||
ErrInsufficientFundsForTransfer = errors.New("insufficient funds for transfer")
|
ErrInsufficientFundsForTransfer = errors.New("insufficient funds for transfer")
|
||||||
|
|
|
||||||
|
|
@ -44,6 +44,7 @@ func NewEVMBlockContext(header *types.Header, chain ChainContext, author *common
|
||||||
baseFee *big.Int
|
baseFee *big.Int
|
||||||
blobBaseFee *big.Int
|
blobBaseFee *big.Int
|
||||||
random *common.Hash
|
random *common.Hash
|
||||||
|
slotNum uint64
|
||||||
)
|
)
|
||||||
|
|
||||||
// If we don't have an explicit author (i.e. not mining), extract from the header
|
// If we don't have an explicit author (i.e. not mining), extract from the header
|
||||||
|
|
@ -61,6 +62,10 @@ func NewEVMBlockContext(header *types.Header, chain ChainContext, author *common
|
||||||
if header.Difficulty.Sign() == 0 {
|
if header.Difficulty.Sign() == 0 {
|
||||||
random = &header.MixDigest
|
random = &header.MixDigest
|
||||||
}
|
}
|
||||||
|
if header.SlotNumber != nil {
|
||||||
|
slotNum = *header.SlotNumber
|
||||||
|
}
|
||||||
|
|
||||||
return vm.BlockContext{
|
return vm.BlockContext{
|
||||||
CanTransfer: CanTransfer,
|
CanTransfer: CanTransfer,
|
||||||
Transfer: Transfer,
|
Transfer: Transfer,
|
||||||
|
|
@ -73,6 +78,7 @@ func NewEVMBlockContext(header *types.Header, chain ChainContext, author *common
|
||||||
BlobBaseFee: blobBaseFee,
|
BlobBaseFee: blobBaseFee,
|
||||||
GasLimit: header.GasLimit,
|
GasLimit: header.GasLimit,
|
||||||
Random: random,
|
Random: random,
|
||||||
|
SlotNum: slotNum,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
||||||
|
|
@ -21,39 +21,87 @@ import (
|
||||||
"math"
|
"math"
|
||||||
)
|
)
|
||||||
|
|
||||||
// GasPool tracks the amount of gas available during execution of the transactions
|
// GasPool tracks the amount of gas available for transaction execution
|
||||||
// in a block. The zero value is a pool with zero gas available.
|
// within a block, along with the cumulative gas consumed.
|
||||||
type GasPool uint64
|
type GasPool struct {
|
||||||
|
remaining uint64
|
||||||
|
initial uint64
|
||||||
|
cumulativeUsed uint64
|
||||||
|
}
|
||||||
|
|
||||||
// AddGas makes gas available for execution.
|
// NewGasPool initializes the gasPool with the given amount.
|
||||||
func (gp *GasPool) AddGas(amount uint64) *GasPool {
|
func NewGasPool(amount uint64) *GasPool {
|
||||||
if uint64(*gp) > math.MaxUint64-amount {
|
return &GasPool{
|
||||||
panic("gas pool pushed above uint64")
|
remaining: amount,
|
||||||
|
initial: amount,
|
||||||
}
|
}
|
||||||
*(*uint64)(gp) += amount
|
|
||||||
return gp
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// SubGas deducts the given amount from the pool if enough gas is
|
// SubGas deducts the given amount from the pool if enough gas is
|
||||||
// available and returns an error otherwise.
|
// available and returns an error otherwise.
|
||||||
func (gp *GasPool) SubGas(amount uint64) error {
|
func (gp *GasPool) SubGas(amount uint64) error {
|
||||||
if uint64(*gp) < amount {
|
if gp.remaining < amount {
|
||||||
return ErrGasLimitReached
|
return ErrGasLimitReached
|
||||||
}
|
}
|
||||||
*(*uint64)(gp) -= amount
|
gp.remaining -= amount
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// ReturnGas adds the refunded gas back to the pool and updates
|
||||||
|
// the cumulative gas usage accordingly.
|
||||||
|
func (gp *GasPool) ReturnGas(returned uint64, gasUsed uint64) error {
|
||||||
|
if gp.remaining > math.MaxUint64-returned {
|
||||||
|
return fmt.Errorf("%w: remaining: %d, returned: %d", ErrGasLimitOverflow, gp.remaining, returned)
|
||||||
|
}
|
||||||
|
// The returned gas calculation differs across forks.
|
||||||
|
//
|
||||||
|
// - Pre-Amsterdam:
|
||||||
|
// returned = purchased - remaining (refund included)
|
||||||
|
//
|
||||||
|
// - Post-Amsterdam:
|
||||||
|
// returned = purchased - gasUsed (refund excluded)
|
||||||
|
gp.remaining += returned
|
||||||
|
|
||||||
|
// gasUsed = max(txGasUsed - gasRefund, calldataFloorGasCost)
|
||||||
|
// regardless of Amsterdam is activated or not.
|
||||||
|
gp.cumulativeUsed += gasUsed
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// Gas returns the amount of gas remaining in the pool.
|
// Gas returns the amount of gas remaining in the pool.
|
||||||
func (gp *GasPool) Gas() uint64 {
|
func (gp *GasPool) Gas() uint64 {
|
||||||
return uint64(*gp)
|
return gp.remaining
|
||||||
}
|
}
|
||||||
|
|
||||||
// SetGas sets the amount of gas with the provided number.
|
// CumulativeUsed returns the amount of cumulative consumed gas (refunded included).
|
||||||
func (gp *GasPool) SetGas(gas uint64) {
|
func (gp *GasPool) CumulativeUsed() uint64 {
|
||||||
*(*uint64)(gp) = gas
|
return gp.cumulativeUsed
|
||||||
|
}
|
||||||
|
|
||||||
|
// Used returns the amount of consumed gas.
|
||||||
|
func (gp *GasPool) Used() uint64 {
|
||||||
|
if gp.initial < gp.remaining {
|
||||||
|
panic("gas used underflow")
|
||||||
|
}
|
||||||
|
return gp.initial - gp.remaining
|
||||||
|
}
|
||||||
|
|
||||||
|
// Snapshot returns the deep-copied object as the snapshot.
|
||||||
|
func (gp *GasPool) Snapshot() *GasPool {
|
||||||
|
return &GasPool{
|
||||||
|
initial: gp.initial,
|
||||||
|
remaining: gp.remaining,
|
||||||
|
cumulativeUsed: gp.cumulativeUsed,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Set sets the content of gasPool with the provided one.
|
||||||
|
func (gp *GasPool) Set(other *GasPool) {
|
||||||
|
gp.initial = other.initial
|
||||||
|
gp.remaining = other.remaining
|
||||||
|
gp.cumulativeUsed = other.cumulativeUsed
|
||||||
}
|
}
|
||||||
|
|
||||||
func (gp *GasPool) String() string {
|
func (gp *GasPool) String() string {
|
||||||
return fmt.Sprintf("%d", *gp)
|
return fmt.Sprintf("initial: %d, remaining: %d, cumulative used: %d", gp.initial, gp.remaining, gp.cumulativeUsed)
|
||||||
}
|
}
|
||||||
|
|
|
||||||
|
|
@ -34,6 +34,7 @@ func (g Genesis) MarshalJSON() ([]byte, error) {
|
||||||
BaseFee *math.HexOrDecimal256 `json:"baseFeePerGas"`
|
BaseFee *math.HexOrDecimal256 `json:"baseFeePerGas"`
|
||||||
ExcessBlobGas *math.HexOrDecimal64 `json:"excessBlobGas"`
|
ExcessBlobGas *math.HexOrDecimal64 `json:"excessBlobGas"`
|
||||||
BlobGasUsed *math.HexOrDecimal64 `json:"blobGasUsed"`
|
BlobGasUsed *math.HexOrDecimal64 `json:"blobGasUsed"`
|
||||||
|
SlotNumber *uint64 `json:"slotNumber"`
|
||||||
}
|
}
|
||||||
var enc Genesis
|
var enc Genesis
|
||||||
enc.Config = g.Config
|
enc.Config = g.Config
|
||||||
|
|
@ -56,6 +57,7 @@ func (g Genesis) MarshalJSON() ([]byte, error) {
|
||||||
enc.BaseFee = (*math.HexOrDecimal256)(g.BaseFee)
|
enc.BaseFee = (*math.HexOrDecimal256)(g.BaseFee)
|
||||||
enc.ExcessBlobGas = (*math.HexOrDecimal64)(g.ExcessBlobGas)
|
enc.ExcessBlobGas = (*math.HexOrDecimal64)(g.ExcessBlobGas)
|
||||||
enc.BlobGasUsed = (*math.HexOrDecimal64)(g.BlobGasUsed)
|
enc.BlobGasUsed = (*math.HexOrDecimal64)(g.BlobGasUsed)
|
||||||
|
enc.SlotNumber = g.SlotNumber
|
||||||
return json.Marshal(&enc)
|
return json.Marshal(&enc)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
@ -77,6 +79,7 @@ func (g *Genesis) UnmarshalJSON(input []byte) error {
|
||||||
BaseFee *math.HexOrDecimal256 `json:"baseFeePerGas"`
|
BaseFee *math.HexOrDecimal256 `json:"baseFeePerGas"`
|
||||||
ExcessBlobGas *math.HexOrDecimal64 `json:"excessBlobGas"`
|
ExcessBlobGas *math.HexOrDecimal64 `json:"excessBlobGas"`
|
||||||
BlobGasUsed *math.HexOrDecimal64 `json:"blobGasUsed"`
|
BlobGasUsed *math.HexOrDecimal64 `json:"blobGasUsed"`
|
||||||
|
SlotNumber *uint64 `json:"slotNumber"`
|
||||||
}
|
}
|
||||||
var dec Genesis
|
var dec Genesis
|
||||||
if err := json.Unmarshal(input, &dec); err != nil {
|
if err := json.Unmarshal(input, &dec); err != nil {
|
||||||
|
|
@ -133,5 +136,8 @@ func (g *Genesis) UnmarshalJSON(input []byte) error {
|
||||||
if dec.BlobGasUsed != nil {
|
if dec.BlobGasUsed != nil {
|
||||||
g.BlobGasUsed = (*uint64)(dec.BlobGasUsed)
|
g.BlobGasUsed = (*uint64)(dec.BlobGasUsed)
|
||||||
}
|
}
|
||||||
|
if dec.SlotNumber != nil {
|
||||||
|
g.SlotNumber = dec.SlotNumber
|
||||||
|
}
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
|
||||||
|
|
@ -73,6 +73,7 @@ type Genesis struct {
|
||||||
BaseFee *big.Int `json:"baseFeePerGas"` // EIP-1559
|
BaseFee *big.Int `json:"baseFeePerGas"` // EIP-1559
|
||||||
ExcessBlobGas *uint64 `json:"excessBlobGas"` // EIP-4844
|
ExcessBlobGas *uint64 `json:"excessBlobGas"` // EIP-4844
|
||||||
BlobGasUsed *uint64 `json:"blobGasUsed"` // EIP-4844
|
BlobGasUsed *uint64 `json:"blobGasUsed"` // EIP-4844
|
||||||
|
SlotNumber *uint64 `json:"slotNumber"` // EIP-7843
|
||||||
}
|
}
|
||||||
|
|
||||||
// copy copies the genesis.
|
// copy copies the genesis.
|
||||||
|
|
@ -122,6 +123,7 @@ func ReadGenesis(db ethdb.Database) (*Genesis, error) {
|
||||||
genesis.BaseFee = genesisHeader.BaseFee
|
genesis.BaseFee = genesisHeader.BaseFee
|
||||||
genesis.ExcessBlobGas = genesisHeader.ExcessBlobGas
|
genesis.ExcessBlobGas = genesisHeader.ExcessBlobGas
|
||||||
genesis.BlobGasUsed = genesisHeader.BlobGasUsed
|
genesis.BlobGasUsed = genesisHeader.BlobGasUsed
|
||||||
|
genesis.SlotNumber = genesisHeader.SlotNumber
|
||||||
|
|
||||||
return &genesis, nil
|
return &genesis, nil
|
||||||
}
|
}
|
||||||
|
|
@ -547,6 +549,12 @@ func (g *Genesis) toBlockWithRoot(root common.Hash) *types.Block {
|
||||||
if conf.IsPrague(num, g.Timestamp) {
|
if conf.IsPrague(num, g.Timestamp) {
|
||||||
head.RequestsHash = &types.EmptyRequestsHash
|
head.RequestsHash = &types.EmptyRequestsHash
|
||||||
}
|
}
|
||||||
|
if conf.IsAmsterdam(num, g.Timestamp) {
|
||||||
|
head.SlotNumber = g.SlotNumber
|
||||||
|
if head.SlotNumber == nil {
|
||||||
|
head.SlotNumber = new(uint64)
|
||||||
|
}
|
||||||
|
}
|
||||||
}
|
}
|
||||||
return types.NewBlock(head, &types.Body{Withdrawals: withdrawals}, nil, trie.NewStackTrie(nil))
|
return types.NewBlock(head, &types.Body{Withdrawals: withdrawals}, nil, trie.NewStackTrie(nil))
|
||||||
}
|
}
|
||||||
|
|
|
||||||
|
|
@ -424,13 +424,7 @@ func WriteBodyRLP(db ethdb.KeyValueWriter, hash common.Hash, number uint64, rlp
|
||||||
// HasBody verifies the existence of a block body corresponding to the hash.
|
// HasBody verifies the existence of a block body corresponding to the hash.
|
||||||
func HasBody(db ethdb.Reader, hash common.Hash, number uint64) bool {
|
func HasBody(db ethdb.Reader, hash common.Hash, number uint64) bool {
|
||||||
if isCanon(db, number, hash) {
|
if isCanon(db, number, hash) {
|
||||||
// Block is in ancient store, but bodies can be pruned.
|
return true
|
||||||
// Check if the block number is above the pruning tail.
|
|
||||||
tail, _ := db.Tail()
|
|
||||||
if number >= tail {
|
|
||||||
return true
|
|
||||||
}
|
|
||||||
return false
|
|
||||||
}
|
}
|
||||||
if has, err := db.Has(blockBodyKey(number, hash)); !has || err != nil {
|
if has, err := db.Has(blockBodyKey(number, hash)); !has || err != nil {
|
||||||
return false
|
return false
|
||||||
|
|
@ -472,13 +466,7 @@ func DeleteBody(db ethdb.KeyValueWriter, hash common.Hash, number uint64) {
|
||||||
// to a block.
|
// to a block.
|
||||||
func HasReceipts(db ethdb.Reader, hash common.Hash, number uint64) bool {
|
func HasReceipts(db ethdb.Reader, hash common.Hash, number uint64) bool {
|
||||||
if isCanon(db, number, hash) {
|
if isCanon(db, number, hash) {
|
||||||
// Block is in ancient store, but receipts can be pruned.
|
return true
|
||||||
// Check if the block number is above the pruning tail.
|
|
||||||
tail, _ := db.Tail()
|
|
||||||
if number >= tail {
|
|
||||||
return true
|
|
||||||
}
|
|
||||||
return false
|
|
||||||
}
|
}
|
||||||
if has, err := db.Has(blockReceiptsKey(number, hash)); !has || err != nil {
|
if has, err := db.Has(blockReceiptsKey(number, hash)); !has || err != nil {
|
||||||
return false
|
return false
|
||||||
|
|
|
||||||
|
|
@ -35,6 +35,7 @@ import (
|
||||||
"github.com/ethereum/go-ethereum/crypto"
|
"github.com/ethereum/go-ethereum/crypto"
|
||||||
"github.com/ethereum/go-ethereum/ethdb"
|
"github.com/ethereum/go-ethereum/ethdb"
|
||||||
"github.com/ethereum/go-ethereum/ethdb/memorydb"
|
"github.com/ethereum/go-ethereum/ethdb/memorydb"
|
||||||
|
"github.com/ethereum/go-ethereum/internal/tablewriter"
|
||||||
"github.com/ethereum/go-ethereum/log"
|
"github.com/ethereum/go-ethereum/log"
|
||||||
"golang.org/x/sync/errgroup"
|
"golang.org/x/sync/errgroup"
|
||||||
)
|
)
|
||||||
|
|
@ -663,7 +664,7 @@ func InspectDatabase(db ethdb.Database, keyPrefix, keyStart []byte) error {
|
||||||
total.Add(uint64(ancient.size()))
|
total.Add(uint64(ancient.size()))
|
||||||
}
|
}
|
||||||
|
|
||||||
table := NewTableWriter(os.Stdout)
|
table := tablewriter.NewWriter(os.Stdout)
|
||||||
table.SetHeader([]string{"Database", "Category", "Size", "Items"})
|
table.SetHeader([]string{"Database", "Category", "Size", "Items"})
|
||||||
table.SetFooter([]string{"", "Total", common.StorageSize(total.Load()).String(), fmt.Sprintf("%d", count.Load())})
|
table.SetFooter([]string{"", "Total", common.StorageSize(total.Load()).String(), fmt.Sprintf("%d", count.Load())})
|
||||||
table.AppendBulk(stats)
|
table.AppendBulk(stats)
|
||||||
|
|
|
||||||
|
|
@ -253,6 +253,11 @@ func (b *tableBatch) Reset() {
|
||||||
b.batch.Reset()
|
b.batch.Reset()
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Close closes the batch and releases all associated resources.
|
||||||
|
func (b *tableBatch) Close() {
|
||||||
|
b.batch.Close()
|
||||||
|
}
|
||||||
|
|
||||||
// tableReplayer is a wrapper around a batch replayer which truncates
|
// tableReplayer is a wrapper around a batch replayer which truncates
|
||||||
// the added prefix.
|
// the added prefix.
|
||||||
type tableReplayer struct {
|
type tableReplayer struct {
|
||||||
|
|
|
||||||
|
|
@ -195,6 +195,19 @@ func (s *stateObject) GetCommittedState(key common.Hash) common.Hash {
|
||||||
// have been handles via pendingStorage above.
|
// have been handles via pendingStorage above.
|
||||||
// 2) we don't have new values, and can deliver empty response back
|
// 2) we don't have new values, and can deliver empty response back
|
||||||
if _, destructed := s.db.stateObjectsDestruct[s.address]; destructed {
|
if _, destructed := s.db.stateObjectsDestruct[s.address]; destructed {
|
||||||
|
// Invoke the reader regardless and discard the returned value.
|
||||||
|
// The returned value may not be empty, as it could belong to a
|
||||||
|
// self-destructed contract.
|
||||||
|
//
|
||||||
|
// The read operation is still essential for correctly building
|
||||||
|
// the block-level access list.
|
||||||
|
//
|
||||||
|
// TODO(rjl493456442) the reader interface can be extended with
|
||||||
|
// Touch, recording the read access without the actual disk load.
|
||||||
|
_, err := s.db.reader.Storage(s.address, key)
|
||||||
|
if err != nil {
|
||||||
|
s.db.setError(err)
|
||||||
|
}
|
||||||
s.originStorage[key] = common.Hash{} // track the empty slot as origin value
|
s.originStorage[key] = common.Hash{} // track the empty slot as origin value
|
||||||
return common.Hash{}
|
return common.Hash{}
|
||||||
}
|
}
|
||||||
|
|
|
||||||
|
|
@ -1342,6 +1342,7 @@ func (s *StateDB) commitAndFlush(block uint64, deleteEmptyObjects bool, noStorag
|
||||||
if err := batch.Write(); err != nil {
|
if err := batch.Write(); err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
|
batch.Close()
|
||||||
}
|
}
|
||||||
if !ret.empty() {
|
if !ret.empty() {
|
||||||
// If snapshotting is enabled, update the snapshot tree with this new version
|
// If snapshotting is enabled, update the snapshot tree with this new version
|
||||||
|
|
|
||||||
|
|
@ -107,7 +107,7 @@ func (p *statePrefetcher) Prefetch(block *types.Block, statedb *state.StateDB, c
|
||||||
|
|
||||||
// We attempt to apply a transaction. The goal is not to execute
|
// We attempt to apply a transaction. The goal is not to execute
|
||||||
// the transaction successfully, rather to warm up touched data slots.
|
// the transaction successfully, rather to warm up touched data slots.
|
||||||
if _, err := ApplyMessage(evm, msg, new(GasPool).AddGas(block.GasLimit())); err != nil {
|
if _, err := ApplyMessage(evm, msg, nil); err != nil {
|
||||||
fails.Add(1)
|
fails.Add(1)
|
||||||
return nil // Ugh, something went horribly wrong, bail out
|
return nil // Ugh, something went horribly wrong, bail out
|
||||||
}
|
}
|
||||||
|
|
|
||||||
|
|
@ -63,14 +63,12 @@ func (p *StateProcessor) Process(ctx context.Context, block *types.Block, stated
|
||||||
var (
|
var (
|
||||||
config = p.chainConfig()
|
config = p.chainConfig()
|
||||||
receipts types.Receipts
|
receipts types.Receipts
|
||||||
usedGas = new(uint64)
|
|
||||||
header = block.Header()
|
header = block.Header()
|
||||||
blockHash = block.Hash()
|
blockHash = block.Hash()
|
||||||
blockNumber = block.Number()
|
blockNumber = block.Number()
|
||||||
allLogs []*types.Log
|
allLogs []*types.Log
|
||||||
gp = new(GasPool).AddGas(block.GasLimit())
|
gp = NewGasPool(block.GasLimit())
|
||||||
)
|
)
|
||||||
|
|
||||||
var tracingStateDB = vm.StateDB(statedb)
|
var tracingStateDB = vm.StateDB(statedb)
|
||||||
if hooks := cfg.Tracer; hooks != nil {
|
if hooks := cfg.Tracer; hooks != nil {
|
||||||
tracingStateDB = state.NewHookedState(statedb, hooks)
|
tracingStateDB = state.NewHookedState(statedb, hooks)
|
||||||
|
|
@ -107,13 +105,15 @@ func (p *StateProcessor) Process(ctx context.Context, block *types.Block, stated
|
||||||
telemetry.StringAttribute("tx.hash", tx.Hash().Hex()),
|
telemetry.StringAttribute("tx.hash", tx.Hash().Hex()),
|
||||||
telemetry.Int64Attribute("tx.index", int64(i)),
|
telemetry.Int64Attribute("tx.index", int64(i)),
|
||||||
)
|
)
|
||||||
receipt, err := ApplyTransactionWithEVM(msg, gp, statedb, blockNumber, blockHash, context.Time, tx, usedGas, evm)
|
|
||||||
spanEnd(&err)
|
receipt, err := ApplyTransactionWithEVM(msg, gp, statedb, blockNumber, blockHash, context.Time, tx, evm)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, fmt.Errorf("could not apply tx %d [%v]: %w", i, tx.Hash().Hex(), err)
|
return nil, fmt.Errorf("could not apply tx %d [%v]: %w", i, tx.Hash().Hex(), err)
|
||||||
}
|
}
|
||||||
receipts = append(receipts, receipt)
|
receipts = append(receipts, receipt)
|
||||||
allLogs = append(allLogs, receipt.Logs...)
|
allLogs = append(allLogs, receipt.Logs...)
|
||||||
|
|
||||||
|
spanEnd(&err)
|
||||||
}
|
}
|
||||||
requests, err := postExecution(ctx, config, block, allLogs, evm)
|
requests, err := postExecution(ctx, config, block, allLogs, evm)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
|
|
@ -127,7 +127,7 @@ func (p *StateProcessor) Process(ctx context.Context, block *types.Block, stated
|
||||||
Receipts: receipts,
|
Receipts: receipts,
|
||||||
Requests: requests,
|
Requests: requests,
|
||||||
Logs: allLogs,
|
Logs: allLogs,
|
||||||
GasUsed: *usedGas,
|
GasUsed: gp.Used(),
|
||||||
}, nil
|
}, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
@ -159,7 +159,7 @@ func postExecution(ctx context.Context, config *params.ChainConfig, block *types
|
||||||
// ApplyTransactionWithEVM attempts to apply a transaction to the given state database
|
// ApplyTransactionWithEVM attempts to apply a transaction to the given state database
|
||||||
// and uses the input parameters for its environment similar to ApplyTransaction. However,
|
// and uses the input parameters for its environment similar to ApplyTransaction. However,
|
||||||
// this method takes an already created EVM instance as input.
|
// this method takes an already created EVM instance as input.
|
||||||
func ApplyTransactionWithEVM(msg *Message, gp *GasPool, statedb *state.StateDB, blockNumber *big.Int, blockHash common.Hash, blockTime uint64, tx *types.Transaction, usedGas *uint64, evm *vm.EVM) (receipt *types.Receipt, err error) {
|
func ApplyTransactionWithEVM(msg *Message, gp *GasPool, statedb *state.StateDB, blockNumber *big.Int, blockHash common.Hash, blockTime uint64, tx *types.Transaction, evm *vm.EVM) (receipt *types.Receipt, err error) {
|
||||||
if hooks := evm.Config.Tracer; hooks != nil {
|
if hooks := evm.Config.Tracer; hooks != nil {
|
||||||
if hooks.OnTxStart != nil {
|
if hooks.OnTxStart != nil {
|
||||||
hooks.OnTxStart(evm.GetVMContext(), tx, msg.From)
|
hooks.OnTxStart(evm.GetVMContext(), tx, msg.From)
|
||||||
|
|
@ -180,27 +180,31 @@ func ApplyTransactionWithEVM(msg *Message, gp *GasPool, statedb *state.StateDB,
|
||||||
} else {
|
} else {
|
||||||
root = statedb.IntermediateRoot(evm.ChainConfig().IsEIP158(blockNumber)).Bytes()
|
root = statedb.IntermediateRoot(evm.ChainConfig().IsEIP158(blockNumber)).Bytes()
|
||||||
}
|
}
|
||||||
*usedGas += result.UsedGas
|
|
||||||
|
|
||||||
// Merge the tx-local access event into the "block-local" one, in order to collect
|
// Merge the tx-local access event into the "block-local" one, in order to collect
|
||||||
// all values, so that the witness can be built.
|
// all values, so that the witness can be built.
|
||||||
if statedb.Database().TrieDB().IsVerkle() {
|
if statedb.Database().TrieDB().IsVerkle() {
|
||||||
statedb.AccessEvents().Merge(evm.AccessEvents)
|
statedb.AccessEvents().Merge(evm.AccessEvents)
|
||||||
}
|
}
|
||||||
return MakeReceipt(evm, result, statedb, blockNumber, blockHash, blockTime, tx, *usedGas, root), nil
|
return MakeReceipt(evm, result, statedb, blockNumber, blockHash, blockTime, tx, gp.CumulativeUsed(), root), nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// MakeReceipt generates the receipt object for a transaction given its execution result.
|
// MakeReceipt generates the receipt object for a transaction given its execution result.
|
||||||
func MakeReceipt(evm *vm.EVM, result *ExecutionResult, statedb *state.StateDB, blockNumber *big.Int, blockHash common.Hash, blockTime uint64, tx *types.Transaction, usedGas uint64, root []byte) *types.Receipt {
|
func MakeReceipt(evm *vm.EVM, result *ExecutionResult, statedb *state.StateDB, blockNumber *big.Int, blockHash common.Hash, blockTime uint64, tx *types.Transaction, cumulativeGas uint64, root []byte) *types.Receipt {
|
||||||
// Create a new receipt for the transaction, storing the intermediate root and gas used
|
// Create a new receipt for the transaction, storing the intermediate root
|
||||||
// by the tx.
|
// and gas used by the tx.
|
||||||
receipt := &types.Receipt{Type: tx.Type(), PostState: root, CumulativeGasUsed: usedGas}
|
//
|
||||||
|
// The cumulative gas used equals the sum of gasUsed across all preceding
|
||||||
|
// txs with refunded gas deducted.
|
||||||
|
receipt := &types.Receipt{Type: tx.Type(), PostState: root, CumulativeGasUsed: cumulativeGas}
|
||||||
if result.Failed() {
|
if result.Failed() {
|
||||||
receipt.Status = types.ReceiptStatusFailed
|
receipt.Status = types.ReceiptStatusFailed
|
||||||
} else {
|
} else {
|
||||||
receipt.Status = types.ReceiptStatusSuccessful
|
receipt.Status = types.ReceiptStatusSuccessful
|
||||||
}
|
}
|
||||||
receipt.TxHash = tx.Hash()
|
receipt.TxHash = tx.Hash()
|
||||||
|
|
||||||
|
// GasUsed = max(tx_gas_used - gas_refund, calldata_floor_gas_cost), unchanged
|
||||||
|
// in the Amsterdam fork.
|
||||||
receipt.GasUsed = result.UsedGas
|
receipt.GasUsed = result.UsedGas
|
||||||
|
|
||||||
if tx.Type() == types.BlobTxType {
|
if tx.Type() == types.BlobTxType {
|
||||||
|
|
@ -224,15 +228,15 @@ func MakeReceipt(evm *vm.EVM, result *ExecutionResult, statedb *state.StateDB, b
|
||||||
|
|
||||||
// ApplyTransaction attempts to apply a transaction to the given state database
|
// ApplyTransaction attempts to apply a transaction to the given state database
|
||||||
// and uses the input parameters for its environment. It returns the receipt
|
// and uses the input parameters for its environment. It returns the receipt
|
||||||
// for the transaction, gas used and an error if the transaction failed,
|
// for the transaction and an error if the transaction failed,
|
||||||
// indicating the block was invalid.
|
// indicating the block was invalid.
|
||||||
func ApplyTransaction(evm *vm.EVM, gp *GasPool, statedb *state.StateDB, header *types.Header, tx *types.Transaction, usedGas *uint64) (*types.Receipt, error) {
|
func ApplyTransaction(evm *vm.EVM, gp *GasPool, statedb *state.StateDB, header *types.Header, tx *types.Transaction) (*types.Receipt, error) {
|
||||||
msg, err := TransactionToMessage(tx, types.MakeSigner(evm.ChainConfig(), header.Number, header.Time), header.BaseFee)
|
msg, err := TransactionToMessage(tx, types.MakeSigner(evm.ChainConfig(), header.Number, header.Time), header.BaseFee)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
// Create a new context to be used in the EVM environment
|
// Create a new context to be used in the EVM environment
|
||||||
return ApplyTransactionWithEVM(msg, gp, statedb, header.Number, header.Hash(), header.Time, tx, usedGas, evm)
|
return ApplyTransactionWithEVM(msg, gp, statedb, header.Number, header.Hash(), header.Time, tx, evm)
|
||||||
}
|
}
|
||||||
|
|
||||||
// ProcessBeaconBlockRoot applies the EIP-4788 system call to the beacon block root
|
// ProcessBeaconBlockRoot applies the EIP-4788 system call to the beacon block root
|
||||||
|
|
|
||||||
|
|
@ -422,6 +422,9 @@ func GenerateBadBlock(parent *types.Block, engine consensus.Engine, txs types.Tr
|
||||||
beaconRoot := common.HexToHash("0xbeac00")
|
beaconRoot := common.HexToHash("0xbeac00")
|
||||||
header.ParentBeaconRoot = &beaconRoot
|
header.ParentBeaconRoot = &beaconRoot
|
||||||
}
|
}
|
||||||
|
if config.IsAmsterdam(header.Number, header.Time) {
|
||||||
|
header.SlotNumber = new(uint64)
|
||||||
|
}
|
||||||
// Assemble and return the final block for sealing
|
// Assemble and return the final block for sealing
|
||||||
body := &types.Body{Transactions: txs}
|
body := &types.Body{Transactions: txs}
|
||||||
if config.IsShanghai(header.Number, header.Time) {
|
if config.IsShanghai(header.Number, header.Time) {
|
||||||
|
|
|
||||||
|
|
@ -34,7 +34,7 @@ import (
|
||||||
// ExecutionResult includes all output after executing given evm
|
// ExecutionResult includes all output after executing given evm
|
||||||
// message no matter the execution itself is successful or not.
|
// message no matter the execution itself is successful or not.
|
||||||
type ExecutionResult struct {
|
type ExecutionResult struct {
|
||||||
UsedGas uint64 // Total used gas, not including the refunded gas
|
UsedGas uint64 // Total used gas, refunded gas is deducted
|
||||||
MaxUsedGas uint64 // Maximum gas consumed during execution, excluding gas refunds.
|
MaxUsedGas uint64 // Maximum gas consumed during execution, excluding gas refunds.
|
||||||
Err error // Any error encountered during the execution(listed in core/vm/errors.go)
|
Err error // Any error encountered during the execution(listed in core/vm/errors.go)
|
||||||
ReturnData []byte // Returned data from evm(function result or data supplied with revert opcode)
|
ReturnData []byte // Returned data from evm(function result or data supplied with revert opcode)
|
||||||
|
|
@ -177,9 +177,9 @@ func TransactionToMessage(tx *types.Transaction, s types.Signer, baseFee *big.In
|
||||||
msg := &Message{
|
msg := &Message{
|
||||||
Nonce: tx.Nonce(),
|
Nonce: tx.Nonce(),
|
||||||
GasLimit: tx.Gas(),
|
GasLimit: tx.Gas(),
|
||||||
GasPrice: new(big.Int).Set(tx.GasPrice()),
|
GasPrice: tx.GasPrice(),
|
||||||
GasFeeCap: new(big.Int).Set(tx.GasFeeCap()),
|
GasFeeCap: tx.GasFeeCap(),
|
||||||
GasTipCap: new(big.Int).Set(tx.GasTipCap()),
|
GasTipCap: tx.GasTipCap(),
|
||||||
To: tx.To(),
|
To: tx.To(),
|
||||||
Value: tx.Value(),
|
Value: tx.Value(),
|
||||||
Data: tx.Data(),
|
Data: tx.Data(),
|
||||||
|
|
@ -210,6 +210,11 @@ func TransactionToMessage(tx *types.Transaction, s types.Signer, baseFee *big.In
|
||||||
// indicates a core error meaning that the message would always fail for that particular
|
// indicates a core error meaning that the message would always fail for that particular
|
||||||
// state and would never be accepted within a block.
|
// state and would never be accepted within a block.
|
||||||
func ApplyMessage(evm *vm.EVM, msg *Message, gp *GasPool) (*ExecutionResult, error) {
|
func ApplyMessage(evm *vm.EVM, msg *Message, gp *GasPool) (*ExecutionResult, error) {
|
||||||
|
// Do not panic if the gas pool is nil. This is allowed when executing
|
||||||
|
// a single message via RPC invocation.
|
||||||
|
if gp == nil {
|
||||||
|
gp = NewGasPool(msg.GasLimit)
|
||||||
|
}
|
||||||
evm.SetTxContext(NewEVMTxContext(msg))
|
evm.SetTxContext(NewEVMTxContext(msg))
|
||||||
return newStateTransition(evm, msg, gp).execute()
|
return newStateTransition(evm, msg, gp).execute()
|
||||||
}
|
}
|
||||||
|
|
@ -300,8 +305,8 @@ func (st *stateTransition) buyGas() error {
|
||||||
st.evm.Config.Tracer.OnGasChange(0, st.msg.GasLimit, tracing.GasChangeTxInitialBalance)
|
st.evm.Config.Tracer.OnGasChange(0, st.msg.GasLimit, tracing.GasChangeTxInitialBalance)
|
||||||
}
|
}
|
||||||
st.gasRemaining = st.msg.GasLimit
|
st.gasRemaining = st.msg.GasLimit
|
||||||
|
|
||||||
st.initialGas = st.msg.GasLimit
|
st.initialGas = st.msg.GasLimit
|
||||||
|
|
||||||
mgvalU256, _ := uint256.FromBig(mgval)
|
mgvalU256, _ := uint256.FromBig(mgval)
|
||||||
st.state.SubBalance(st.msg.From, mgvalU256, tracing.BalanceDecreaseGasBuy)
|
st.state.SubBalance(st.msg.From, mgvalU256, tracing.BalanceDecreaseGasBuy)
|
||||||
return nil
|
return nil
|
||||||
|
|
@ -542,8 +547,20 @@ func (st *stateTransition) execute() (*ExecutionResult, error) {
|
||||||
peakGasUsed = floorDataGas
|
peakGasUsed = floorDataGas
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
// Return gas to the user
|
||||||
st.returnGas()
|
st.returnGas()
|
||||||
|
|
||||||
|
// Return gas to the gas pool
|
||||||
|
if rules.IsAmsterdam {
|
||||||
|
// Refund is excluded for returning
|
||||||
|
err = st.gp.ReturnGas(st.initialGas-peakGasUsed, st.gasUsed())
|
||||||
|
} else {
|
||||||
|
// Refund is included for returning
|
||||||
|
err = st.gp.ReturnGas(st.gasRemaining, st.gasUsed())
|
||||||
|
}
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
effectiveTip := msg.GasPrice
|
effectiveTip := msg.GasPrice
|
||||||
if rules.IsLondon {
|
if rules.IsLondon {
|
||||||
effectiveTip = new(big.Int).Sub(msg.GasPrice, st.evm.Context.BaseFee)
|
effectiveTip = new(big.Int).Sub(msg.GasPrice, st.evm.Context.BaseFee)
|
||||||
|
|
@ -564,7 +581,6 @@ func (st *stateTransition) execute() (*ExecutionResult, error) {
|
||||||
st.evm.AccessEvents.AddAccount(st.evm.Context.Coinbase, true, math.MaxUint64)
|
st.evm.AccessEvents.AddAccount(st.evm.Context.Coinbase, true, math.MaxUint64)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
return &ExecutionResult{
|
return &ExecutionResult{
|
||||||
UsedGas: st.gasUsed(),
|
UsedGas: st.gasUsed(),
|
||||||
MaxUsedGas: peakGasUsed,
|
MaxUsedGas: peakGasUsed,
|
||||||
|
|
@ -660,10 +676,6 @@ func (st *stateTransition) returnGas() {
|
||||||
if st.evm.Config.Tracer != nil && st.evm.Config.Tracer.OnGasChange != nil && st.gasRemaining > 0 {
|
if st.evm.Config.Tracer != nil && st.evm.Config.Tracer.OnGasChange != nil && st.gasRemaining > 0 {
|
||||||
st.evm.Config.Tracer.OnGasChange(st.gasRemaining, 0, tracing.GasChangeTxLeftOverReturned)
|
st.evm.Config.Tracer.OnGasChange(st.gasRemaining, 0, tracing.GasChangeTxLeftOverReturned)
|
||||||
}
|
}
|
||||||
|
|
||||||
// Also return remaining gas to the block gas counter so it is
|
|
||||||
// available for the next transaction.
|
|
||||||
st.gp.AddGas(st.gasRemaining)
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// gasUsed returns the amount of gas used up by the state transition.
|
// gasUsed returns the amount of gas used up by the state transition.
|
||||||
|
|
|
||||||
|
|
@ -27,6 +27,7 @@ import (
|
||||||
"github.com/ethereum/go-ethereum/common"
|
"github.com/ethereum/go-ethereum/common"
|
||||||
"github.com/ethereum/go-ethereum/log"
|
"github.com/ethereum/go-ethereum/log"
|
||||||
"github.com/ethereum/go-ethereum/metrics"
|
"github.com/ethereum/go-ethereum/metrics"
|
||||||
|
"github.com/ethereum/go-ethereum/trie"
|
||||||
)
|
)
|
||||||
|
|
||||||
var accountTrieLeavesAtDepth [16]*metrics.Counter
|
var accountTrieLeavesAtDepth [16]*metrics.Counter
|
||||||
|
|
@ -41,59 +42,68 @@ func init() {
|
||||||
|
|
||||||
// WitnessStats aggregates statistics for account and storage trie accesses.
|
// WitnessStats aggregates statistics for account and storage trie accesses.
|
||||||
type WitnessStats struct {
|
type WitnessStats struct {
|
||||||
accountTrieLeaves [16]int64
|
accountTrie *trie.LevelStats
|
||||||
storageTrieLeaves [16]int64
|
storageTrie *trie.LevelStats
|
||||||
}
|
}
|
||||||
|
|
||||||
// NewWitnessStats creates a new WitnessStats collector.
|
// NewWitnessStats creates a new WitnessStats collector.
|
||||||
func NewWitnessStats() *WitnessStats {
|
func NewWitnessStats() *WitnessStats {
|
||||||
return &WitnessStats{}
|
return &WitnessStats{
|
||||||
|
accountTrie: trie.NewLevelStats(),
|
||||||
|
storageTrie: trie.NewLevelStats(),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func (s *WitnessStats) init() {
|
||||||
|
if s.accountTrie == nil {
|
||||||
|
s.accountTrie = trie.NewLevelStats()
|
||||||
|
}
|
||||||
|
if s.storageTrie == nil {
|
||||||
|
s.storageTrie = trie.NewLevelStats()
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Add records trie access depths from the given node paths.
|
// Add records trie access depths from the given node paths.
|
||||||
// If `owner` is the zero hash, accesses are attributed to the account trie;
|
// If `owner` is the zero hash, accesses are attributed to the account trie;
|
||||||
// otherwise, they are attributed to the storage trie of that account.
|
// otherwise, they are attributed to the storage trie of that account.
|
||||||
func (s *WitnessStats) Add(nodes map[string][]byte, owner common.Hash) {
|
func (s *WitnessStats) Add(nodes map[string][]byte, owner common.Hash) {
|
||||||
// Extract paths from the nodes map
|
s.init()
|
||||||
|
|
||||||
|
// Extract paths from the nodes map.
|
||||||
paths := slices.Collect(maps.Keys(nodes))
|
paths := slices.Collect(maps.Keys(nodes))
|
||||||
sort.Strings(paths)
|
sort.Strings(paths)
|
||||||
|
|
||||||
|
ownerStat := s.accountTrie
|
||||||
|
if owner != (common.Hash{}) {
|
||||||
|
ownerStat = s.storageTrie
|
||||||
|
}
|
||||||
|
|
||||||
for i, path := range paths {
|
for i, path := range paths {
|
||||||
// If current path is a prefix of the next path, it's not a leaf.
|
// If current path is a prefix of the next path, it's not a leaf.
|
||||||
// The last path is always a leaf.
|
// The last path is always a leaf.
|
||||||
if i == len(paths)-1 || !strings.HasPrefix(paths[i+1], paths[i]) {
|
if i == len(paths)-1 || !strings.HasPrefix(paths[i+1], paths[i]) {
|
||||||
depth := len(path)
|
ownerStat.AddLeaf(len(path))
|
||||||
if owner == (common.Hash{}) {
|
|
||||||
if depth >= len(s.accountTrieLeaves) {
|
|
||||||
depth = len(s.accountTrieLeaves) - 1
|
|
||||||
}
|
|
||||||
s.accountTrieLeaves[depth] += 1
|
|
||||||
} else {
|
|
||||||
if depth >= len(s.storageTrieLeaves) {
|
|
||||||
depth = len(s.storageTrieLeaves) - 1
|
|
||||||
}
|
|
||||||
s.storageTrieLeaves[depth] += 1
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// ReportMetrics reports the collected statistics to the global metrics registry.
|
// ReportMetrics reports the collected statistics to the global metrics registry.
|
||||||
func (s *WitnessStats) ReportMetrics(blockNumber uint64) {
|
func (s *WitnessStats) ReportMetrics(blockNumber uint64) {
|
||||||
// Encode the metrics as JSON for easier consumption
|
s.init()
|
||||||
accountLeavesJson, _ := json.Marshal(s.accountTrieLeaves)
|
|
||||||
storageLeavesJson, _ := json.Marshal(s.storageTrieLeaves)
|
|
||||||
|
|
||||||
// Log account trie depth statistics
|
accountTrieLeaves := s.accountTrie.LeafDepths()
|
||||||
log.Info("Account trie depth stats",
|
storageTrieLeaves := s.storageTrie.LeafDepths()
|
||||||
"block", blockNumber,
|
|
||||||
"leavesAtDepth", string(accountLeavesJson))
|
|
||||||
log.Info("Storage trie depth stats",
|
|
||||||
"block", blockNumber,
|
|
||||||
"leavesAtDepth", string(storageLeavesJson))
|
|
||||||
|
|
||||||
for i := 0; i < 16; i++ {
|
// Encode the metrics as JSON for easier consumption.
|
||||||
accountTrieLeavesAtDepth[i].Inc(s.accountTrieLeaves[i])
|
accountLeavesJSON, _ := json.Marshal(accountTrieLeaves)
|
||||||
storageTrieLeavesAtDepth[i].Inc(s.storageTrieLeaves[i])
|
storageLeavesJSON, _ := json.Marshal(storageTrieLeaves)
|
||||||
|
|
||||||
|
// Log account trie depth statistics.
|
||||||
|
log.Info("Account trie depth stats", "block", blockNumber, "leavesAtDepth", string(accountLeavesJSON))
|
||||||
|
log.Info("Storage trie depth stats", "block", blockNumber, "leavesAtDepth", string(storageLeavesJSON))
|
||||||
|
|
||||||
|
for i := 0; i < len(accountTrieLeavesAtDepth); i++ {
|
||||||
|
accountTrieLeavesAtDepth[i].Inc(accountTrieLeaves[i])
|
||||||
|
storageTrieLeavesAtDepth[i].Inc(storageTrieLeaves[i])
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
|
||||||
|
|
@ -17,18 +17,27 @@
|
||||||
package stateless
|
package stateless
|
||||||
|
|
||||||
import (
|
import (
|
||||||
|
"strings"
|
||||||
"testing"
|
"testing"
|
||||||
|
|
||||||
"github.com/ethereum/go-ethereum/common"
|
"github.com/ethereum/go-ethereum/common"
|
||||||
)
|
)
|
||||||
|
|
||||||
|
func expectedLeaves(counts map[int]int64) [16]int64 {
|
||||||
|
var leaves [16]int64
|
||||||
|
for depth, count := range counts {
|
||||||
|
leaves[depth] = count
|
||||||
|
}
|
||||||
|
return leaves
|
||||||
|
}
|
||||||
|
|
||||||
func TestWitnessStatsAdd(t *testing.T) {
|
func TestWitnessStatsAdd(t *testing.T) {
|
||||||
tests := []struct {
|
tests := []struct {
|
||||||
name string
|
name string
|
||||||
nodes map[string][]byte
|
nodes map[string][]byte
|
||||||
owner common.Hash
|
owner common.Hash
|
||||||
expectedAccountLeaves map[int64]int64
|
expectedAccountLeaves map[int]int64
|
||||||
expectedStorageLeaves map[int64]int64
|
expectedStorageLeaves map[int]int64
|
||||||
}{
|
}{
|
||||||
{
|
{
|
||||||
name: "empty nodes",
|
name: "empty nodes",
|
||||||
|
|
@ -41,7 +50,7 @@ func TestWitnessStatsAdd(t *testing.T) {
|
||||||
"": []byte("data"),
|
"": []byte("data"),
|
||||||
},
|
},
|
||||||
owner: common.Hash{},
|
owner: common.Hash{},
|
||||||
expectedAccountLeaves: map[int64]int64{0: 1},
|
expectedAccountLeaves: map[int]int64{0: 1},
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
name: "single account trie leaf",
|
name: "single account trie leaf",
|
||||||
|
|
@ -49,7 +58,7 @@ func TestWitnessStatsAdd(t *testing.T) {
|
||||||
"abc": []byte("data"),
|
"abc": []byte("data"),
|
||||||
},
|
},
|
||||||
owner: common.Hash{},
|
owner: common.Hash{},
|
||||||
expectedAccountLeaves: map[int64]int64{3: 1},
|
expectedAccountLeaves: map[int]int64{3: 1},
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
name: "account trie with internal nodes",
|
name: "account trie with internal nodes",
|
||||||
|
|
@ -59,7 +68,7 @@ func TestWitnessStatsAdd(t *testing.T) {
|
||||||
"abc": []byte("data3"),
|
"abc": []byte("data3"),
|
||||||
},
|
},
|
||||||
owner: common.Hash{},
|
owner: common.Hash{},
|
||||||
expectedAccountLeaves: map[int64]int64{3: 1}, // Only "abc" is a leaf
|
expectedAccountLeaves: map[int]int64{3: 1}, // Only "abc" is a leaf
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
name: "multiple account trie branches",
|
name: "multiple account trie branches",
|
||||||
|
|
@ -72,7 +81,7 @@ func TestWitnessStatsAdd(t *testing.T) {
|
||||||
"bcd": []byte("data6"),
|
"bcd": []byte("data6"),
|
||||||
},
|
},
|
||||||
owner: common.Hash{},
|
owner: common.Hash{},
|
||||||
expectedAccountLeaves: map[int64]int64{3: 2}, // "abc" (3) + "bcd" (3)
|
expectedAccountLeaves: map[int]int64{3: 2}, // "abc" (3) + "bcd" (3)
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
name: "siblings are all leaves",
|
name: "siblings are all leaves",
|
||||||
|
|
@ -82,7 +91,7 @@ func TestWitnessStatsAdd(t *testing.T) {
|
||||||
"ac": []byte("data3"),
|
"ac": []byte("data3"),
|
||||||
},
|
},
|
||||||
owner: common.Hash{},
|
owner: common.Hash{},
|
||||||
expectedAccountLeaves: map[int64]int64{2: 3},
|
expectedAccountLeaves: map[int]int64{2: 3},
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
name: "storage trie leaves",
|
name: "storage trie leaves",
|
||||||
|
|
@ -93,7 +102,7 @@ func TestWitnessStatsAdd(t *testing.T) {
|
||||||
"124": []byte("data4"),
|
"124": []byte("data4"),
|
||||||
},
|
},
|
||||||
owner: common.HexToHash("0x1234"),
|
owner: common.HexToHash("0x1234"),
|
||||||
expectedStorageLeaves: map[int64]int64{3: 2}, // "123" (3) + "124" (3)
|
expectedStorageLeaves: map[int]int64{3: 2}, // "123" (3) + "124" (3)
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
name: "complex trie structure",
|
name: "complex trie structure",
|
||||||
|
|
@ -109,7 +118,7 @@ func TestWitnessStatsAdd(t *testing.T) {
|
||||||
"3": []byte("data9"),
|
"3": []byte("data9"),
|
||||||
},
|
},
|
||||||
owner: common.Hash{},
|
owner: common.Hash{},
|
||||||
expectedAccountLeaves: map[int64]int64{1: 1, 3: 4}, // "123"(3) + "124"(3) + "234"(3) + "235"(3) + "3"(1)
|
expectedAccountLeaves: map[int]int64{1: 1, 3: 4}, // "123"(3) + "124"(3) + "234"(3) + "235"(3) + "3"(1)
|
||||||
},
|
},
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
@ -118,32 +127,59 @@ func TestWitnessStatsAdd(t *testing.T) {
|
||||||
stats := NewWitnessStats()
|
stats := NewWitnessStats()
|
||||||
stats.Add(tt.nodes, tt.owner)
|
stats.Add(tt.nodes, tt.owner)
|
||||||
|
|
||||||
var expectedAccountTrieLeaves [16]int64
|
if got, want := stats.accountTrie.LeafDepths(), expectedLeaves(tt.expectedAccountLeaves); got != want {
|
||||||
for depth, count := range tt.expectedAccountLeaves {
|
t.Errorf("account trie leaves = %v, want %v", got, want)
|
||||||
expectedAccountTrieLeaves[depth] = count
|
|
||||||
}
|
}
|
||||||
var expectedStorageTrieLeaves [16]int64
|
if got, want := stats.storageTrie.LeafDepths(), expectedLeaves(tt.expectedStorageLeaves); got != want {
|
||||||
for depth, count := range tt.expectedStorageLeaves {
|
t.Errorf("storage trie leaves = %v, want %v", got, want)
|
||||||
expectedStorageTrieLeaves[depth] = count
|
|
||||||
}
|
|
||||||
|
|
||||||
// Check account trie depth
|
|
||||||
if stats.accountTrieLeaves != expectedAccountTrieLeaves {
|
|
||||||
t.Errorf("Account trie total depth = %v, want %v", stats.accountTrieLeaves, expectedAccountTrieLeaves)
|
|
||||||
}
|
|
||||||
|
|
||||||
// Check storage trie depth
|
|
||||||
if stats.storageTrieLeaves != expectedStorageTrieLeaves {
|
|
||||||
t.Errorf("Storage trie total depth = %v, want %v", stats.storageTrieLeaves, expectedStorageTrieLeaves)
|
|
||||||
}
|
}
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func TestWitnessStatsStorageTrieAggregation(t *testing.T) {
|
||||||
|
stats := NewWitnessStats()
|
||||||
|
ownerA := common.HexToHash("0xa")
|
||||||
|
ownerB := common.HexToHash("0xb")
|
||||||
|
|
||||||
|
stats.Add(map[string][]byte{
|
||||||
|
"a": []byte("data1"),
|
||||||
|
"ab": []byte("data2"),
|
||||||
|
"abc": []byte("data3"),
|
||||||
|
}, ownerA)
|
||||||
|
stats.Add(map[string][]byte{
|
||||||
|
"xy": []byte("data4"),
|
||||||
|
}, ownerA)
|
||||||
|
stats.Add(map[string][]byte{
|
||||||
|
"1": []byte("data5"),
|
||||||
|
"12": []byte("data6"),
|
||||||
|
"123": []byte("data7"),
|
||||||
|
"124": []byte("data8"),
|
||||||
|
}, ownerB)
|
||||||
|
|
||||||
|
if got, want := stats.storageTrie.LeafDepths(), expectedLeaves(map[int]int64{2: 1, 3: 3}); got != want {
|
||||||
|
t.Errorf("storage leaves = %v, want %v", got, want)
|
||||||
|
}
|
||||||
|
if got, want := stats.accountTrie.LeafDepths(), expectedLeaves(nil); got != want {
|
||||||
|
t.Errorf("account leaves = %v, want %v", got, want)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestWitnessStatsPanicsOnDeepLeaf(t *testing.T) {
|
||||||
|
stats := NewWitnessStats()
|
||||||
|
|
||||||
|
defer func() {
|
||||||
|
if r := recover(); r == nil {
|
||||||
|
t.Fatal("expected panic for depth >= 16")
|
||||||
|
}
|
||||||
|
}()
|
||||||
|
stats.Add(map[string][]byte{strings.Repeat("a", 16): []byte("data")}, common.Hash{})
|
||||||
|
}
|
||||||
|
|
||||||
func TestWitnessStatsMinMax(t *testing.T) {
|
func TestWitnessStatsMinMax(t *testing.T) {
|
||||||
stats := NewWitnessStats()
|
stats := NewWitnessStats()
|
||||||
|
|
||||||
// Add some account trie nodes with varying depths
|
// Add some account trie nodes with varying depths.
|
||||||
stats.Add(map[string][]byte{
|
stats.Add(map[string][]byte{
|
||||||
"a": []byte("data1"),
|
"a": []byte("data1"),
|
||||||
"ab": []byte("data2"),
|
"ab": []byte("data2"),
|
||||||
|
|
@ -152,21 +188,21 @@ func TestWitnessStatsMinMax(t *testing.T) {
|
||||||
"abcde": []byte("data5"),
|
"abcde": []byte("data5"),
|
||||||
}, common.Hash{})
|
}, common.Hash{})
|
||||||
|
|
||||||
// Only "abcde" is a leaf (depth 5)
|
// Only "abcde" is a leaf (depth 5).
|
||||||
for i, v := range stats.accountTrieLeaves {
|
for i, v := range stats.accountTrie.LeafDepths() {
|
||||||
if v != 0 && i != 5 {
|
if v != 0 && i != 5 {
|
||||||
t.Errorf("leaf found at invalid depth %d", i)
|
t.Errorf("leaf found at invalid depth %d", i)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Add more leaves with different depths
|
// Add more leaves with different depths.
|
||||||
stats.Add(map[string][]byte{
|
stats.Add(map[string][]byte{
|
||||||
"x": []byte("data6"),
|
"x": []byte("data6"),
|
||||||
"yz": []byte("data7"),
|
"yz": []byte("data7"),
|
||||||
}, common.Hash{})
|
}, common.Hash{})
|
||||||
|
|
||||||
// Now we have leaves at depths 1, 2, and 5
|
// Now we have leaves at depths 1, 2, and 5.
|
||||||
for i, v := range stats.accountTrieLeaves {
|
for i, v := range stats.accountTrie.LeafDepths() {
|
||||||
if v != 0 && (i != 5 && i != 2 && i != 1) {
|
if v != 0 && (i != 5 && i != 2 && i != 1) {
|
||||||
t.Errorf("leaf found at invalid depth %d", i)
|
t.Errorf("leaf found at invalid depth %d", i)
|
||||||
}
|
}
|
||||||
|
|
@ -176,7 +212,7 @@ func TestWitnessStatsMinMax(t *testing.T) {
|
||||||
func TestWitnessStatsAverage(t *testing.T) {
|
func TestWitnessStatsAverage(t *testing.T) {
|
||||||
stats := NewWitnessStats()
|
stats := NewWitnessStats()
|
||||||
|
|
||||||
// Add nodes that will create leaves at depths 2, 3, and 4
|
// Add nodes that will create leaves at depths 2, 3, and 4.
|
||||||
stats.Add(map[string][]byte{
|
stats.Add(map[string][]byte{
|
||||||
"aa": []byte("data1"),
|
"aa": []byte("data1"),
|
||||||
"bb": []byte("data2"),
|
"bb": []byte("data2"),
|
||||||
|
|
@ -184,22 +220,22 @@ func TestWitnessStatsAverage(t *testing.T) {
|
||||||
"dddd": []byte("data4"),
|
"dddd": []byte("data4"),
|
||||||
}, common.Hash{})
|
}, common.Hash{})
|
||||||
|
|
||||||
// All are leaves: 2 + 2 + 3 + 4 = 11 total, 4 samples
|
// All are leaves: 2 + 2 + 3 + 4 = 11 total, 4 samples.
|
||||||
expectedAvg := int64(11) / int64(4)
|
expectedAvg := int64(11) / int64(4)
|
||||||
var actualAvg, totalSamples int64
|
var actualAvg, totalSamples int64
|
||||||
for i, c := range stats.accountTrieLeaves {
|
for i, c := range stats.accountTrie.LeafDepths() {
|
||||||
actualAvg += c * int64(i)
|
actualAvg += c * int64(i)
|
||||||
totalSamples += c
|
totalSamples += c
|
||||||
}
|
}
|
||||||
actualAvg = actualAvg / totalSamples
|
actualAvg = actualAvg / totalSamples
|
||||||
|
|
||||||
if actualAvg != expectedAvg {
|
if actualAvg != expectedAvg {
|
||||||
t.Errorf("Account trie average depth = %d, want %d", actualAvg, expectedAvg)
|
t.Errorf("account trie average depth = %d, want %d", actualAvg, expectedAvg)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func BenchmarkWitnessStatsAdd(b *testing.B) {
|
func BenchmarkWitnessStatsAdd(b *testing.B) {
|
||||||
// Create a realistic trie node structure
|
// Create a realistic trie node structure.
|
||||||
nodes := make(map[string][]byte)
|
nodes := make(map[string][]byte)
|
||||||
for i := 0; i < 100; i++ {
|
for i := 0; i < 100; i++ {
|
||||||
base := string(rune('a' + i%26))
|
base := string(rune('a' + i%26))
|
||||||
|
|
|
||||||
|
|
@ -358,7 +358,7 @@ const (
|
||||||
// this generates an increase in gas. There is at most one of such gas change per transaction.
|
// this generates an increase in gas. There is at most one of such gas change per transaction.
|
||||||
GasChangeTxRefunds GasChangeReason = 3
|
GasChangeTxRefunds GasChangeReason = 3
|
||||||
// GasChangeTxLeftOverReturned is the amount of gas left over at the end of transaction's execution that will be returned
|
// GasChangeTxLeftOverReturned is the amount of gas left over at the end of transaction's execution that will be returned
|
||||||
// to the chain. This change will always be a negative change as we "drain" left over gas towards 0. If there was no gas
|
// to the account. This change will always be a negative change as we "drain" left over gas towards 0. If there was no gas
|
||||||
// left at the end of execution, no such even will be emitted. The returned gas's value in Wei is returned to caller.
|
// left at the end of execution, no such even will be emitted. The returned gas's value in Wei is returned to caller.
|
||||||
// There is at most one of such gas change per transaction.
|
// There is at most one of such gas change per transaction.
|
||||||
GasChangeTxLeftOverReturned GasChangeReason = 4
|
GasChangeTxLeftOverReturned GasChangeReason = 4
|
||||||
|
|
|
||||||
|
|
@ -104,6 +104,16 @@ const (
|
||||||
// maxGappedTxs is the maximum number of gapped transactions kept overall.
|
// maxGappedTxs is the maximum number of gapped transactions kept overall.
|
||||||
// This is a safety limit to avoid DoS vectors.
|
// This is a safety limit to avoid DoS vectors.
|
||||||
maxGapped = 128
|
maxGapped = 128
|
||||||
|
|
||||||
|
// notifyThreshold is the eviction priority threshold above which a transaction
|
||||||
|
// is considered close enough to being includable to be announced to peers.
|
||||||
|
// Setting this to zero will disable announcements for anyting not immediately
|
||||||
|
// includable. Setting it to -1 allows transactions that are close to being
|
||||||
|
// includable, maybe already in the next block if fees go down, to be announced.
|
||||||
|
|
||||||
|
// Note, this threshold is in the abstract eviction priority space, so its
|
||||||
|
// meaning depends on the current basefee/blobfee and the transaction's fees.
|
||||||
|
announceThreshold = -1
|
||||||
)
|
)
|
||||||
|
|
||||||
// blobTxMeta is the minimal subset of types.BlobTx necessary to validate and
|
// blobTxMeta is the minimal subset of types.BlobTx necessary to validate and
|
||||||
|
|
@ -115,6 +125,8 @@ type blobTxMeta struct {
|
||||||
vhashes []common.Hash // Blob versioned hashes to maintain the lookup table
|
vhashes []common.Hash // Blob versioned hashes to maintain the lookup table
|
||||||
version byte // Blob transaction version to determine proof type
|
version byte // Blob transaction version to determine proof type
|
||||||
|
|
||||||
|
announced bool // Whether the tx has been announced to listeners
|
||||||
|
|
||||||
id uint64 // Storage ID in the pool's persistent store
|
id uint64 // Storage ID in the pool's persistent store
|
||||||
storageSize uint32 // Byte size in the pool's persistent store
|
storageSize uint32 // Byte size in the pool's persistent store
|
||||||
size uint64 // RLP-encoded size of transaction including the attached blob
|
size uint64 // RLP-encoded size of transaction including the attached blob
|
||||||
|
|
@ -159,7 +171,7 @@ func newBlobTxMeta(id uint64, size uint64, storageSize uint32, tx *types.Transac
|
||||||
blobGas: tx.BlobGas(),
|
blobGas: tx.BlobGas(),
|
||||||
}
|
}
|
||||||
meta.basefeeJumps = dynamicFeeJumps(meta.execFeeCap)
|
meta.basefeeJumps = dynamicFeeJumps(meta.execFeeCap)
|
||||||
meta.blobfeeJumps = dynamicFeeJumps(meta.blobFeeCap)
|
meta.blobfeeJumps = dynamicBlobFeeJumps(meta.blobFeeCap)
|
||||||
|
|
||||||
return meta
|
return meta
|
||||||
}
|
}
|
||||||
|
|
@ -210,6 +222,14 @@ func newBlobTxMeta(id uint64, size uint64, storageSize uint32, tx *types.Transac
|
||||||
// via a normal transaction. It should nonetheless be high enough to support
|
// via a normal transaction. It should nonetheless be high enough to support
|
||||||
// resurrecting reorged transactions. Perhaps 4-16.
|
// resurrecting reorged transactions. Perhaps 4-16.
|
||||||
//
|
//
|
||||||
|
// - It is not the role of the blobpool to serve as a storage for limit orders
|
||||||
|
// below market: blob transactions with fee caps way below base fee or blob fee.
|
||||||
|
// Therefore, the propagation of blob transactions that are far from being
|
||||||
|
// includable is suppressed. The pool will only announce blob transactions that
|
||||||
|
// are close to being includable (based on the current fees and the transaction's
|
||||||
|
// fee caps), and will delay the announcement of blob transactions that are far
|
||||||
|
// from being includable until base fee and/or blob fee is reduced.
|
||||||
|
//
|
||||||
// - Local txs are meaningless. Mining pools historically used local transactions
|
// - Local txs are meaningless. Mining pools historically used local transactions
|
||||||
// for payouts or for backdoor deals. With 1559 in place, the basefee usually
|
// for payouts or for backdoor deals. With 1559 in place, the basefee usually
|
||||||
// dominates the final price, so 0 or non-0 tip doesn't change much. Blob txs
|
// dominates the final price, so 0 or non-0 tip doesn't change much. Blob txs
|
||||||
|
|
@ -281,47 +301,54 @@ func newBlobTxMeta(id uint64, size uint64, storageSize uint32, tx *types.Transac
|
||||||
// solve after every block.
|
// solve after every block.
|
||||||
//
|
//
|
||||||
// - The first observation is that comparing 1559 base fees or 4844 blob fees
|
// - The first observation is that comparing 1559 base fees or 4844 blob fees
|
||||||
// needs to happen in the context of their dynamism. Since these fees jump
|
// needs to happen in the context of their dynamism. Since base fees are
|
||||||
// up or down in ~1.125 multipliers (at max) across blocks, comparing fees
|
// adjusted continuously and fluctuate, and we want to optimize for effective
|
||||||
// in two transactions should be based on log1.125(fee) to eliminate noise.
|
// miner fees, it is better to disregard small base fee cap differences.
|
||||||
|
// Instead of considering the exact fee cap values, we should group
|
||||||
|
// transactions into buckets based on fee cap values, allowing us to use
|
||||||
|
// the miner tip meaningfully as a splitter inside a bucket.
|
||||||
//
|
//
|
||||||
// - The second observation is that the basefee and blobfee move independently,
|
// To create these buckets, rather than looking at the absolute fee
|
||||||
// so there's no way to split mixed txs on their own (A has higher base fee,
|
// differences, the useful metric is the max time it can take to exceed the
|
||||||
// B has higher blob fee). Rather than look at the absolute fees, the useful
|
// transaction's fee caps. Base fee changes are multiplicative, so we use a
|
||||||
// metric is the max time it can take to exceed the transaction's fee caps.
|
// logarithmic scale. Fees jumps up or down in ~1.125 multipliers at max
|
||||||
|
// across blocks, so we use log1.125(fee) and rounding to eliminate noise.
|
||||||
// Specifically, we're interested in the number of jumps needed to go from
|
// Specifically, we're interested in the number of jumps needed to go from
|
||||||
// the current fee to the transaction's cap:
|
// the current fee to the transaction's cap:
|
||||||
//
|
//
|
||||||
// jumps = log1.125(txfee) - log1.125(basefee)
|
// jumps = floor(log1.125(txfee) - log1.125(basefee))
|
||||||
//
|
//
|
||||||
// - The third observation is that the base fee tends to hover around rather
|
// For blob fees, EIP-7892 changed the ratio of target to max blobs, and
|
||||||
// than swing wildly. The number of jumps needed from the current fee starts
|
// with that also the maximum blob fee decrease in a slot from 1.125 to
|
||||||
// to get less relevant the higher it is. To remove the noise here too, the
|
// approx 1.17. therefore, we use:
|
||||||
// pool will use log(jumps) as the delta for comparing transactions.
|
|
||||||
//
|
//
|
||||||
// delta = sign(jumps) * log(abs(jumps))
|
// blobfeeJumps = floor(log1.17(txBlobfee) - log1.17(blobfee))
|
||||||
//
|
//
|
||||||
// - To establish a total order, we need to reduce the dimensionality of the
|
// - The second observation is that when ranking executable blob txs, it
|
||||||
|
// does not make sense to grant a later eviction priority to txs with high
|
||||||
|
// fee caps since it could enable pool wars. As such, any positive priority
|
||||||
|
// will be grouped together.
|
||||||
|
//
|
||||||
|
// priority = min(jumps, 0)
|
||||||
|
//
|
||||||
|
// - The third observation is that the basefee and blobfee move independently,
|
||||||
|
// so there's no way to split mixed txs on their own (A has higher base fee,
|
||||||
|
// B has higher blob fee).
|
||||||
|
//
|
||||||
|
// To establish a total order, we need to reduce the dimensionality of the
|
||||||
// two base fees (log jumps) to a single value. The interesting aspect from
|
// two base fees (log jumps) to a single value. The interesting aspect from
|
||||||
// the pool's perspective is how fast will a tx get executable (fees going
|
// the pool's perspective is how fast will a tx get executable (fees going
|
||||||
// down, crossing the smaller negative jump counter) or non-executable (fees
|
// down, crossing the smaller negative jump counter) or non-executable (fees
|
||||||
// going up, crossing the smaller positive jump counter). As such, the pool
|
// going up, crossing the smaller positive jump counter). As such, the pool
|
||||||
// cares only about the min of the two delta values for eviction priority.
|
// cares only about the min of the two delta values for eviction priority.
|
||||||
//
|
//
|
||||||
// priority = min(deltaBasefee, deltaBlobfee)
|
// priority = min(deltaBasefee, deltaBlobfee, 0)
|
||||||
//
|
//
|
||||||
// - The above very aggressive dimensionality and noise reduction should result
|
// - The above very aggressive dimensionality and noise reduction should result
|
||||||
// in transaction being grouped into a small number of buckets, the further
|
// in transaction being grouped into a small number of buckets, the further
|
||||||
// the fees the larger the buckets. This is good because it allows us to use
|
// the fees the larger the buckets. This is good because it allows us to use
|
||||||
// the miner tip meaningfully as a splitter.
|
// the miner tip meaningfully as a splitter.
|
||||||
//
|
//
|
||||||
// - For the scenario where the pool does not contain non-executable blob txs
|
|
||||||
// anymore, it does not make sense to grant a later eviction priority to txs
|
|
||||||
// with high fee caps since it could enable pool wars. As such, any positive
|
|
||||||
// priority will be grouped together.
|
|
||||||
//
|
|
||||||
// priority = min(deltaBasefee, deltaBlobfee, 0)
|
|
||||||
//
|
|
||||||
// Optimisation tradeoffs:
|
// Optimisation tradeoffs:
|
||||||
//
|
//
|
||||||
// - Eviction relies on 3 fee minimums per account (exec tip, exec cap and blob
|
// - Eviction relies on 3 fee minimums per account (exec tip, exec cap and blob
|
||||||
|
|
@ -470,6 +497,20 @@ func (p *BlobPool) Init(gasTip uint64, head *types.Header, reserver txpool.Reser
|
||||||
}
|
}
|
||||||
p.evict = newPriceHeap(basefee, blobfee, p.index)
|
p.evict = newPriceHeap(basefee, blobfee, p.index)
|
||||||
|
|
||||||
|
// Guess what was announced. This is needed because we don't want to
|
||||||
|
// participate in the diffusion of transactions where inclusion is blocked by
|
||||||
|
// a low base fee transaction. Since we don't persist that info, the best
|
||||||
|
// we can do is to assume that anything that could have been announced
|
||||||
|
// at current prices, actually was.
|
||||||
|
for addr := range p.index {
|
||||||
|
for _, tx := range p.index[addr] {
|
||||||
|
tx.announced = p.isAnnouncable(tx)
|
||||||
|
if !tx.announced {
|
||||||
|
break
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
// Pool initialized, attach the blob limbo to it to track blobs included
|
// Pool initialized, attach the blob limbo to it to track blobs included
|
||||||
// recently but not yet finalized
|
// recently but not yet finalized
|
||||||
p.limbo, err = newLimbo(p.chain.Config(), limbodir)
|
p.limbo, err = newLimbo(p.chain.Config(), limbodir)
|
||||||
|
|
@ -517,6 +558,7 @@ func (p *BlobPool) Close() error {
|
||||||
|
|
||||||
// parseTransaction is a callback method on pool creation that gets called for
|
// parseTransaction is a callback method on pool creation that gets called for
|
||||||
// each transaction on disk to create the in-memory metadata index.
|
// each transaction on disk to create the in-memory metadata index.
|
||||||
|
// Announced state is not initialized here, it needs to be iniitalized seprately.
|
||||||
func (p *BlobPool) parseTransaction(id uint64, size uint32, blob []byte) error {
|
func (p *BlobPool) parseTransaction(id uint64, size uint32, blob []byte) error {
|
||||||
tx := new(types.Transaction)
|
tx := new(types.Transaction)
|
||||||
if err := rlp.DecodeBytes(blob, tx); err != nil {
|
if err := rlp.DecodeBytes(blob, tx); err != nil {
|
||||||
|
|
@ -893,6 +935,37 @@ func (p *BlobPool) Reset(oldHead, newHead *types.Header) {
|
||||||
}
|
}
|
||||||
p.evict.reinit(basefee, blobfee, false)
|
p.evict.reinit(basefee, blobfee, false)
|
||||||
|
|
||||||
|
// Announce transactions that became announcable due to fee changes
|
||||||
|
var announcable []*types.Transaction
|
||||||
|
for addr, txs := range p.index {
|
||||||
|
for i, meta := range txs {
|
||||||
|
if !meta.announced && (i == 0 || txs[i-1].announced) && p.isAnnouncable(meta) {
|
||||||
|
// Load the full transaction and strip the sidecar before announcing
|
||||||
|
// TODO: this is a bit ugly, as we have everything needed in meta already
|
||||||
|
data, err := p.store.Get(meta.id)
|
||||||
|
// Technically, we are supposed to set announced only if Get is successful.
|
||||||
|
// However, Get failing here indicates a more serious issue (data loss),
|
||||||
|
// so we set announced anyway to avoid repeated attempts.
|
||||||
|
meta.announced = true
|
||||||
|
if err != nil {
|
||||||
|
log.Error("Blobs missing for announcable transaction", "from", addr, "nonce", meta.nonce, "id", meta.id, "err", err)
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
var tx types.Transaction
|
||||||
|
if err = rlp.DecodeBytes(data, &tx); err != nil {
|
||||||
|
log.Error("Blobs corrupted for announcable transaction", "from", addr, "nonce", meta.nonce, "id", meta.id, "err", err)
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
announcable = append(announcable, tx.WithoutBlobTxSidecar())
|
||||||
|
log.Trace("Blob transaction now announcable", "from", addr, "nonce", meta.nonce, "id", meta.id, "hash", tx.Hash())
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if len(announcable) > 0 {
|
||||||
|
p.discoverFeed.Send(core.NewTxsEvent{Txs: announcable})
|
||||||
|
}
|
||||||
|
|
||||||
|
// Update the basefee and blobfee metrics
|
||||||
basefeeGauge.Update(int64(basefee.Uint64()))
|
basefeeGauge.Update(int64(basefee.Uint64()))
|
||||||
blobfeeGauge.Update(int64(blobfee.Uint64()))
|
blobfeeGauge.Update(int64(blobfee.Uint64()))
|
||||||
p.updateStorageMetrics()
|
p.updateStorageMetrics()
|
||||||
|
|
@ -1476,17 +1549,12 @@ func (p *BlobPool) AvailableBlobs(vhashes []common.Hash) int {
|
||||||
// Add inserts a set of blob transactions into the pool if they pass validation (both
|
// Add inserts a set of blob transactions into the pool if they pass validation (both
|
||||||
// consensus validity and pool restrictions).
|
// consensus validity and pool restrictions).
|
||||||
func (p *BlobPool) Add(txs []*types.Transaction, sync bool) []error {
|
func (p *BlobPool) Add(txs []*types.Transaction, sync bool) []error {
|
||||||
var (
|
errs := make([]error, len(txs))
|
||||||
errs = make([]error, len(txs))
|
|
||||||
adds = make([]*types.Transaction, 0, len(txs))
|
|
||||||
)
|
|
||||||
for i, tx := range txs {
|
for i, tx := range txs {
|
||||||
if errs[i] = p.ValidateTxBasics(tx); errs[i] != nil {
|
if errs[i] = p.ValidateTxBasics(tx); errs[i] != nil {
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
if errs[i] = p.add(tx); errs[i] == nil {
|
errs[i] = p.add(tx)
|
||||||
adds = append(adds, tx.WithoutBlobTxSidecar())
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
return errs
|
return errs
|
||||||
}
|
}
|
||||||
|
|
@ -1678,9 +1746,15 @@ func (p *BlobPool) addLocked(tx *types.Transaction, checkGapped bool) (err error
|
||||||
|
|
||||||
addValidMeter.Mark(1)
|
addValidMeter.Mark(1)
|
||||||
|
|
||||||
// Notify all listeners of the new arrival
|
// Transaction was addded successfully, but we only announce if it is (close to being)
|
||||||
p.discoverFeed.Send(core.NewTxsEvent{Txs: []*types.Transaction{tx.WithoutBlobTxSidecar()}})
|
// includable and the previous one was already announced.
|
||||||
p.insertFeed.Send(core.NewTxsEvent{Txs: []*types.Transaction{tx.WithoutBlobTxSidecar()}})
|
if p.isAnnouncable(meta) && (meta.nonce == next || (len(txs) > 1 && txs[offset-1].announced)) {
|
||||||
|
meta.announced = true
|
||||||
|
p.discoverFeed.Send(core.NewTxsEvent{Txs: []*types.Transaction{tx.WithoutBlobTxSidecar()}})
|
||||||
|
p.insertFeed.Send(core.NewTxsEvent{Txs: []*types.Transaction{tx.WithoutBlobTxSidecar()}})
|
||||||
|
} else {
|
||||||
|
log.Trace("Blob transaction not announcable yet", "hash", tx.Hash(), "nonce", tx.Nonce())
|
||||||
|
}
|
||||||
|
|
||||||
//check the gapped queue for this account and try to promote
|
//check the gapped queue for this account and try to promote
|
||||||
if gtxs, ok := p.gapped[from]; checkGapped && ok && len(gtxs) > 0 {
|
if gtxs, ok := p.gapped[from]; checkGapped && ok && len(gtxs) > 0 {
|
||||||
|
|
@ -2004,6 +2078,12 @@ func (p *BlobPool) evictGapped() {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// isAnnouncable checks whether a transaction is announcable based on its
|
||||||
|
// fee parameters and announceThreshold.
|
||||||
|
func (p *BlobPool) isAnnouncable(meta *blobTxMeta) bool {
|
||||||
|
return evictionPriority(p.evict.basefeeJumps, meta.basefeeJumps, p.evict.blobfeeJumps, meta.blobfeeJumps) >= announceThreshold
|
||||||
|
}
|
||||||
|
|
||||||
// Stats retrieves the current pool stats, namely the number of pending and the
|
// Stats retrieves the current pool stats, namely the number of pending and the
|
||||||
// number of queued (non-executable) transactions.
|
// number of queued (non-executable) transactions.
|
||||||
func (p *BlobPool) Stats() (int, int) {
|
func (p *BlobPool) Stats() (int, int) {
|
||||||
|
|
|
||||||
|
|
@ -830,8 +830,8 @@ func TestOpenIndex(t *testing.T) {
|
||||||
//blobfeeJumps = []float64{34.023, 35.570, 36.879, 29.686, 26.243, 20.358} // log 1.125 (blob fee cap)
|
//blobfeeJumps = []float64{34.023, 35.570, 36.879, 29.686, 26.243, 20.358} // log 1.125 (blob fee cap)
|
||||||
|
|
||||||
evictExecTipCaps = []uint64{10, 10, 5, 5, 1, 1}
|
evictExecTipCaps = []uint64{10, 10, 5, 5, 1, 1}
|
||||||
evictExecFeeJumps = []float64{39.098, 38.204, 38.204, 19.549, 19.549, 19.549} // min(log 1.125 (exec fee cap))
|
evictExecFeeJumps = []float64{39.098, 38.204, 38.204, 19.549, 19.549, 19.549} // min(log 1.125 (exec fee cap))
|
||||||
evictBlobFeeJumps = []float64{34.023, 34.023, 34.023, 29.686, 26.243, 20.358} // min(log 1.125 (blob fee cap))
|
evictBlobFeeJumps = []float64{25.517256, 25.517256, 25.517256, 22.264502, 19.682646, 15.268934} // min(log 1.17 (blob fee cap))
|
||||||
|
|
||||||
totalSpent = uint256.NewInt(21000*(100+90+200+10+80+300) + blobSize*(55+66+77+33+22+11) + 100*6) // 21000 gas x price + 128KB x blobprice + value
|
totalSpent = uint256.NewInt(21000*(100+90+200+10+80+300) + blobSize*(55+66+77+33+22+11) + 100*6) // 21000 gas x price + 128KB x blobprice + value
|
||||||
)
|
)
|
||||||
|
|
@ -1751,8 +1751,8 @@ func TestAdd(t *testing.T) {
|
||||||
// Create a blob pool out of the pre-seeded dats
|
// Create a blob pool out of the pre-seeded dats
|
||||||
chain := &testBlockChain{
|
chain := &testBlockChain{
|
||||||
config: params.MainnetChainConfig,
|
config: params.MainnetChainConfig,
|
||||||
basefee: uint256.NewInt(1050),
|
basefee: uint256.NewInt(1),
|
||||||
blobfee: uint256.NewInt(105),
|
blobfee: uint256.NewInt(1),
|
||||||
statedb: statedb,
|
statedb: statedb,
|
||||||
}
|
}
|
||||||
pool := New(Config{Datadir: storage}, chain, nil)
|
pool := New(Config{Datadir: storage}, chain, nil)
|
||||||
|
|
|
||||||
|
|
@ -67,7 +67,7 @@ func newPriceHeap(basefee *uint256.Int, blobfee *uint256.Int, index map[common.A
|
||||||
func (h *evictHeap) reinit(basefee *uint256.Int, blobfee *uint256.Int, force bool) {
|
func (h *evictHeap) reinit(basefee *uint256.Int, blobfee *uint256.Int, force bool) {
|
||||||
// If the update is mostly the same as the old, don't sort pointlessly
|
// If the update is mostly the same as the old, don't sort pointlessly
|
||||||
basefeeJumps := dynamicFeeJumps(basefee)
|
basefeeJumps := dynamicFeeJumps(basefee)
|
||||||
blobfeeJumps := dynamicFeeJumps(blobfee)
|
blobfeeJumps := dynamicBlobFeeJumps(blobfee)
|
||||||
|
|
||||||
if !force && math.Abs(h.basefeeJumps-basefeeJumps) < 0.01 && math.Abs(h.blobfeeJumps-blobfeeJumps) < 0.01 { // TODO(karalabe): 0.01 enough, maybe should be smaller? Maybe this optimization is moot?
|
if !force && math.Abs(h.basefeeJumps-basefeeJumps) < 0.01 && math.Abs(h.blobfeeJumps-blobfeeJumps) < 0.01 { // TODO(karalabe): 0.01 enough, maybe should be smaller? Maybe this optimization is moot?
|
||||||
return
|
return
|
||||||
|
|
@ -95,13 +95,7 @@ func (h *evictHeap) Less(i, j int) bool {
|
||||||
lastJ := txsJ[len(txsJ)-1]
|
lastJ := txsJ[len(txsJ)-1]
|
||||||
|
|
||||||
prioI := evictionPriority(h.basefeeJumps, lastI.evictionExecFeeJumps, h.blobfeeJumps, lastI.evictionBlobFeeJumps)
|
prioI := evictionPriority(h.basefeeJumps, lastI.evictionExecFeeJumps, h.blobfeeJumps, lastI.evictionBlobFeeJumps)
|
||||||
if prioI > 0 {
|
|
||||||
prioI = 0
|
|
||||||
}
|
|
||||||
prioJ := evictionPriority(h.basefeeJumps, lastJ.evictionExecFeeJumps, h.blobfeeJumps, lastJ.evictionBlobFeeJumps)
|
prioJ := evictionPriority(h.basefeeJumps, lastJ.evictionExecFeeJumps, h.blobfeeJumps, lastJ.evictionBlobFeeJumps)
|
||||||
if prioJ > 0 {
|
|
||||||
prioJ = 0
|
|
||||||
}
|
|
||||||
if prioI == prioJ {
|
if prioI == prioJ {
|
||||||
return lastI.evictionExecTip.Lt(lastJ.evictionExecTip)
|
return lastI.evictionExecTip.Lt(lastJ.evictionExecTip)
|
||||||
}
|
}
|
||||||
|
|
|
||||||
|
|
@ -109,22 +109,22 @@ func TestPriceHeapSorting(t *testing.T) {
|
||||||
order: []int{3, 2, 1, 0, 4, 5, 6},
|
order: []int{3, 2, 1, 0, 4, 5, 6},
|
||||||
},
|
},
|
||||||
// If both basefee and blobfee is specified, sort by the larger distance
|
// If both basefee and blobfee is specified, sort by the larger distance
|
||||||
// of the two from the current network conditions, splitting same (loglog)
|
// of the two from the current network conditions, splitting same
|
||||||
// ones via the tip.
|
// ones via the tip.
|
||||||
//
|
//
|
||||||
// Basefee: 1000
|
// Basefee: 1000 , jumps: 888, 790, 702, 624
|
||||||
// Blobfee: 100
|
// Blobfee: 100 , jumps: 85, 73, 62, 53
|
||||||
//
|
//
|
||||||
// Tx #0: (800, 80) - 2 jumps below both => priority -1
|
// Tx #0: (800, 80) - 2 jumps below both => priority -2
|
||||||
// Tx #1: (630, 63) - 4 jumps below both => priority -2
|
// Tx #1: (630, 55) - 4 jumps below both => priority -4
|
||||||
// Tx #2: (800, 63) - 2 jumps below basefee, 4 jumps below blobfee => priority -2 (blob penalty dominates)
|
// Tx #2: (800, 55) - 2 jumps below basefee, 4 jumps below blobfee => priority -4 (blob penalty dominates)
|
||||||
// Tx #3: (630, 80) - 4 jumps below basefee, 2 jumps below blobfee => priority -2 (base penalty dominates)
|
// Tx #3: (630, 80) - 4 jumps below basefee, 2 jumps below blobfee => priority -4 (base penalty dominates)
|
||||||
//
|
//
|
||||||
// Txs 1, 2, 3 share the same priority, split via tip, prefer 0 as the best
|
// Txs 1, 2, 3 share the same priority, split via tip, prefer 0 as the best
|
||||||
{
|
{
|
||||||
execTips: []uint64{1, 2, 3, 4},
|
execTips: []uint64{1, 2, 3, 4},
|
||||||
execFees: []uint64{800, 630, 800, 630},
|
execFees: []uint64{800, 630, 800, 630},
|
||||||
blobFees: []uint64{80, 63, 63, 80},
|
blobFees: []uint64{80, 55, 55, 80},
|
||||||
basefee: 1000,
|
basefee: 1000,
|
||||||
blobfee: 100,
|
blobfee: 100,
|
||||||
order: []int{1, 2, 3, 0},
|
order: []int{1, 2, 3, 0},
|
||||||
|
|
@ -142,7 +142,7 @@ func TestPriceHeapSorting(t *testing.T) {
|
||||||
blobFee = uint256.NewInt(tt.blobFees[j])
|
blobFee = uint256.NewInt(tt.blobFees[j])
|
||||||
|
|
||||||
basefeeJumps = dynamicFeeJumps(execFee)
|
basefeeJumps = dynamicFeeJumps(execFee)
|
||||||
blobfeeJumps = dynamicFeeJumps(blobFee)
|
blobfeeJumps = dynamicBlobFeeJumps(blobFee)
|
||||||
)
|
)
|
||||||
index[addr] = []*blobTxMeta{{
|
index[addr] = []*blobTxMeta{{
|
||||||
id: uint64(j),
|
id: uint64(j),
|
||||||
|
|
@ -201,7 +201,7 @@ func benchmarkPriceHeapReinit(b *testing.B, datacap uint64) {
|
||||||
blobFee = uint256.NewInt(rnd.Uint64())
|
blobFee = uint256.NewInt(rnd.Uint64())
|
||||||
|
|
||||||
basefeeJumps = dynamicFeeJumps(execFee)
|
basefeeJumps = dynamicFeeJumps(execFee)
|
||||||
blobfeeJumps = dynamicFeeJumps(blobFee)
|
blobfeeJumps = dynamicBlobFeeJumps(blobFee)
|
||||||
)
|
)
|
||||||
index[addr] = []*blobTxMeta{{
|
index[addr] = []*blobTxMeta{{
|
||||||
id: uint64(i),
|
id: uint64(i),
|
||||||
|
|
@ -277,7 +277,7 @@ func benchmarkPriceHeapOverflow(b *testing.B, datacap uint64) {
|
||||||
blobFee = uint256.NewInt(rnd.Uint64())
|
blobFee = uint256.NewInt(rnd.Uint64())
|
||||||
|
|
||||||
basefeeJumps = dynamicFeeJumps(execFee)
|
basefeeJumps = dynamicFeeJumps(execFee)
|
||||||
blobfeeJumps = dynamicFeeJumps(blobFee)
|
blobfeeJumps = dynamicBlobFeeJumps(blobFee)
|
||||||
)
|
)
|
||||||
index[addr] = []*blobTxMeta{{
|
index[addr] = []*blobTxMeta{{
|
||||||
id: uint64(i),
|
id: uint64(i),
|
||||||
|
|
@ -308,7 +308,7 @@ func benchmarkPriceHeapOverflow(b *testing.B, datacap uint64) {
|
||||||
blobFee = uint256.NewInt(rnd.Uint64())
|
blobFee = uint256.NewInt(rnd.Uint64())
|
||||||
|
|
||||||
basefeeJumps = dynamicFeeJumps(execFee)
|
basefeeJumps = dynamicFeeJumps(execFee)
|
||||||
blobfeeJumps = dynamicFeeJumps(blobFee)
|
blobfeeJumps = dynamicBlobFeeJumps(blobFee)
|
||||||
)
|
)
|
||||||
metas[i] = &blobTxMeta{
|
metas[i] = &blobTxMeta{
|
||||||
id: uint64(int(blobs) + i),
|
id: uint64(int(blobs) + i),
|
||||||
|
|
|
||||||
|
|
@ -18,7 +18,6 @@ package blobpool
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"math"
|
"math"
|
||||||
"math/bits"
|
|
||||||
|
|
||||||
"github.com/holiman/uint256"
|
"github.com/holiman/uint256"
|
||||||
)
|
)
|
||||||
|
|
@ -26,6 +25,13 @@ import (
|
||||||
// log1_125 is used in the eviction priority calculation.
|
// log1_125 is used in the eviction priority calculation.
|
||||||
var log1_125 = math.Log(1.125)
|
var log1_125 = math.Log(1.125)
|
||||||
|
|
||||||
|
// log1_17 is used in the eviction priority calculation for blob fees.
|
||||||
|
// EIP-7892 (BPO) changed the ratio of target to max blobs, and with that
|
||||||
|
// also the maximum blob fee decrease in a slot from 1.125 to approx 1.17 .
|
||||||
|
// Since we want priorities to approximate time, we should change our log
|
||||||
|
// calculation for blob fees.
|
||||||
|
var log1_17 = log1_125 * 4 / 3
|
||||||
|
|
||||||
// evictionPriority calculates the eviction priority based on the algorithm
|
// evictionPriority calculates the eviction priority based on the algorithm
|
||||||
// described in the BlobPool docs for both fee components.
|
// described in the BlobPool docs for both fee components.
|
||||||
//
|
//
|
||||||
|
|
@ -36,23 +42,20 @@ func evictionPriority(basefeeJumps float64, txBasefeeJumps, blobfeeJumps, txBlob
|
||||||
basefeePriority = evictionPriority1D(basefeeJumps, txBasefeeJumps)
|
basefeePriority = evictionPriority1D(basefeeJumps, txBasefeeJumps)
|
||||||
blobfeePriority = evictionPriority1D(blobfeeJumps, txBlobfeeJumps)
|
blobfeePriority = evictionPriority1D(blobfeeJumps, txBlobfeeJumps)
|
||||||
)
|
)
|
||||||
if basefeePriority < blobfeePriority {
|
return min(0, basefeePriority, blobfeePriority)
|
||||||
return basefeePriority
|
|
||||||
}
|
|
||||||
return blobfeePriority
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// evictionPriority1D calculates the eviction priority based on the algorithm
|
// evictionPriority1D calculates the eviction priority based on the algorithm
|
||||||
// described in the BlobPool docs for a single fee component.
|
// described in the BlobPool docs for a single fee component.
|
||||||
func evictionPriority1D(basefeeJumps float64, txfeeJumps float64) int {
|
func evictionPriority1D(basefeeJumps float64, txfeeJumps float64) int {
|
||||||
jumps := txfeeJumps - basefeeJumps
|
jumps := txfeeJumps - basefeeJumps
|
||||||
if int(jumps) == 0 {
|
if jumps <= 0 {
|
||||||
return 0 // can't log2 0
|
return int(math.Floor(jumps))
|
||||||
}
|
}
|
||||||
if jumps < 0 {
|
// We only use the negative part for ordering. The positive part is only used
|
||||||
return -intLog2(uint(-math.Floor(jumps)))
|
// for threshold comparison (with a negative threshold), so the value is almost
|
||||||
}
|
// irrelevant, as long as it's positive.
|
||||||
return intLog2(uint(math.Ceil(jumps)))
|
return int((math.Ceil(jumps)))
|
||||||
}
|
}
|
||||||
|
|
||||||
// dynamicFeeJumps calculates the log1.125(fee), namely the number of fee jumps
|
// dynamicFeeJumps calculates the log1.125(fee), namely the number of fee jumps
|
||||||
|
|
@ -70,21 +73,9 @@ func dynamicFeeJumps(fee *uint256.Int) float64 {
|
||||||
return math.Log(fee.Float64()) / log1_125
|
return math.Log(fee.Float64()) / log1_125
|
||||||
}
|
}
|
||||||
|
|
||||||
// intLog2 is a helper to calculate the integral part of a log2 of an unsigned
|
func dynamicBlobFeeJumps(fee *uint256.Int) float64 {
|
||||||
// integer. It is a very specific calculation that's not particularly useful in
|
if fee.IsZero() {
|
||||||
// general, but it's what we need here (it's fast).
|
return 0 // can't log2 zero, should never happen outside tests, but don't choke
|
||||||
func intLog2(n uint) int {
|
|
||||||
switch {
|
|
||||||
case n == 0:
|
|
||||||
panic("log2(0) is undefined")
|
|
||||||
|
|
||||||
case n < 2048:
|
|
||||||
return bits.UintSize - bits.LeadingZeros(n) - 1
|
|
||||||
|
|
||||||
default:
|
|
||||||
// The input is log1.125(uint256) = log2(uint256) / log2(1.125). At the
|
|
||||||
// most extreme, log2(uint256) will be a bit below 257, and the constant
|
|
||||||
// log2(1.125) ~= 0.17. The larges input thus is ~257 / ~0.17 ~= ~1511.
|
|
||||||
panic("dynamic fee jump diffs cannot reach this")
|
|
||||||
}
|
}
|
||||||
|
return math.Log(fee.Float64()) / log1_17
|
||||||
}
|
}
|
||||||
|
|
|
||||||
|
|
@ -30,12 +30,12 @@ func TestPriorityCalculation(t *testing.T) {
|
||||||
txfee uint64
|
txfee uint64
|
||||||
result int
|
result int
|
||||||
}{
|
}{
|
||||||
{basefee: 7, txfee: 10, result: 2}, // 3.02 jumps, 4 ceil, 2 log2
|
{basefee: 7, txfee: 10, result: 4}, // 3.02 jumps, 4 ceil
|
||||||
{basefee: 17_200_000_000, txfee: 17_200_000_000, result: 0}, // 0 jumps, special case 0 log2
|
{basefee: 17_200_000_000, txfee: 17_200_000_000, result: 0}, // 0 jumps, special case 0
|
||||||
{basefee: 9_853_941_692, txfee: 11_085_092_510, result: 0}, // 0.99 jumps, 1 ceil, 0 log2
|
{basefee: 9_853_941_692, txfee: 11_085_092_510, result: 1}, // 0.99 jumps, 1 ceil
|
||||||
{basefee: 11_544_106_391, txfee: 10_356_781_100, result: 0}, // -0.92 jumps, -1 floor, 0 log2
|
{basefee: 11_544_106_391, txfee: 10_356_781_100, result: -1}, // -0.92 jumps, -1 floor
|
||||||
{basefee: 17_200_000_000, txfee: 7, result: -7}, // -183.57 jumps, -184 floor, -7 log2
|
{basefee: 17_200_000_000, txfee: 7, result: -184}, // -183.57 jumps, -184 floor
|
||||||
{basefee: 7, txfee: 17_200_000_000, result: 7}, // 183.57 jumps, 184 ceil, 7 log2
|
{basefee: 7, txfee: 17_200_000_000, result: 184}, // 183.57 jumps, 184 ceil
|
||||||
}
|
}
|
||||||
for i, tt := range tests {
|
for i, tt := range tests {
|
||||||
var (
|
var (
|
||||||
|
|
@ -69,7 +69,7 @@ func BenchmarkPriorityCalculation(b *testing.B) {
|
||||||
blobfee := uint256.NewInt(123_456_789_000) // Completely random, no idea what this will be
|
blobfee := uint256.NewInt(123_456_789_000) // Completely random, no idea what this will be
|
||||||
|
|
||||||
basefeeJumps := dynamicFeeJumps(basefee)
|
basefeeJumps := dynamicFeeJumps(basefee)
|
||||||
blobfeeJumps := dynamicFeeJumps(blobfee)
|
blobfeeJumps := dynamicBlobFeeJumps(blobfee)
|
||||||
|
|
||||||
// The transaction's fee cap and blob fee cap are constant across the life
|
// The transaction's fee cap and blob fee cap are constant across the life
|
||||||
// of the transaction, so we can pre-calculate and cache them.
|
// of the transaction, so we can pre-calculate and cache them.
|
||||||
|
|
@ -77,7 +77,7 @@ func BenchmarkPriorityCalculation(b *testing.B) {
|
||||||
txBlobfeeJumps := make([]float64, b.N)
|
txBlobfeeJumps := make([]float64, b.N)
|
||||||
for i := 0; i < b.N; i++ {
|
for i := 0; i < b.N; i++ {
|
||||||
txBasefeeJumps[i] = dynamicFeeJumps(uint256.NewInt(rnd.Uint64()))
|
txBasefeeJumps[i] = dynamicFeeJumps(uint256.NewInt(rnd.Uint64()))
|
||||||
txBlobfeeJumps[i] = dynamicFeeJumps(uint256.NewInt(rnd.Uint64()))
|
txBlobfeeJumps[i] = dynamicBlobFeeJumps(uint256.NewInt(rnd.Uint64()))
|
||||||
}
|
}
|
||||||
b.ResetTimer()
|
b.ResetTimer()
|
||||||
b.ReportAllocs()
|
b.ReportAllocs()
|
||||||
|
|
|
||||||
|
|
@ -68,6 +68,43 @@ func TestListAddVeryExpensive(t *testing.T) {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// TestPriceHeapCmp tests that the price heap comparison function works as intended.
|
||||||
|
// It also tests combinations where the basefee is higher than the gas fee cap, which
|
||||||
|
// are useful to sort in the mempool to support basefee changes.
|
||||||
|
func TestPriceHeapCmp(t *testing.T) {
|
||||||
|
key, _ := crypto.GenerateKey()
|
||||||
|
txs := []*types.Transaction{
|
||||||
|
// nonce, gaslimit, gasfee, gastip
|
||||||
|
dynamicFeeTx(0, 1000, big.NewInt(2), big.NewInt(1), key),
|
||||||
|
dynamicFeeTx(0, 1000, big.NewInt(1), big.NewInt(2), key),
|
||||||
|
dynamicFeeTx(0, 1000, big.NewInt(1), big.NewInt(1), key),
|
||||||
|
dynamicFeeTx(0, 1000, big.NewInt(1), big.NewInt(0), key),
|
||||||
|
}
|
||||||
|
|
||||||
|
// create priceHeap
|
||||||
|
ph := &priceHeap{}
|
||||||
|
|
||||||
|
// now set the basefee on the heap
|
||||||
|
for _, basefee := range []uint64{0, 1, 2, 3} {
|
||||||
|
ph.baseFee = uint256.NewInt(basefee)
|
||||||
|
|
||||||
|
for i := 0; i < len(txs); i++ {
|
||||||
|
for j := 0; j < len(txs); j++ {
|
||||||
|
switch {
|
||||||
|
case i == j:
|
||||||
|
if c := ph.cmp(txs[i], txs[j]); c != 0 {
|
||||||
|
t.Errorf("tx %d should be equal priority to tx %d with basefee %d (cmp=%d)", i, j, basefee, c)
|
||||||
|
}
|
||||||
|
case i < j:
|
||||||
|
if c := ph.cmp(txs[i], txs[j]); c != 1 {
|
||||||
|
t.Errorf("tx %d vs tx %d comparison inconsistent with basefee %d (cmp=%d)", i, j, basefee, c)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
func BenchmarkListAdd(b *testing.B) {
|
func BenchmarkListAdd(b *testing.B) {
|
||||||
// Generate a list of transactions to insert
|
// Generate a list of transactions to insert
|
||||||
key, _ := crypto.GenerateKey()
|
key, _ := crypto.GenerateKey()
|
||||||
|
|
|
||||||
|
|
@ -24,13 +24,6 @@ import (
|
||||||
"github.com/holiman/uint256"
|
"github.com/holiman/uint256"
|
||||||
)
|
)
|
||||||
|
|
||||||
// CodeChange contains the runtime bytecode deployed at an address and the
|
|
||||||
// transaction index where the deployment took place.
|
|
||||||
type CodeChange struct {
|
|
||||||
TxIndex uint16
|
|
||||||
Code []byte `json:"code,omitempty"`
|
|
||||||
}
|
|
||||||
|
|
||||||
// ConstructionAccountAccess contains post-block account state for mutations as well as
|
// ConstructionAccountAccess contains post-block account state for mutations as well as
|
||||||
// all storage keys that were read during execution. It is used when building block
|
// all storage keys that were read during execution. It is used when building block
|
||||||
// access list during execution.
|
// access list during execution.
|
||||||
|
|
@ -55,9 +48,9 @@ type ConstructionAccountAccess struct {
|
||||||
// by tx index.
|
// by tx index.
|
||||||
NonceChanges map[uint16]uint64 `json:"nonceChanges,omitempty"`
|
NonceChanges map[uint16]uint64 `json:"nonceChanges,omitempty"`
|
||||||
|
|
||||||
// CodeChange is only set for contract accounts which were deployed in
|
// CodeChange contains the post-state contract code of an account keyed
|
||||||
// the block.
|
// by tx index.
|
||||||
CodeChange *CodeChange `json:"codeChange,omitempty"`
|
CodeChange map[uint16][]byte `json:"codeChange,omitempty"`
|
||||||
}
|
}
|
||||||
|
|
||||||
// NewConstructionAccountAccess initializes the account access object.
|
// NewConstructionAccountAccess initializes the account access object.
|
||||||
|
|
@ -67,6 +60,7 @@ func NewConstructionAccountAccess() *ConstructionAccountAccess {
|
||||||
StorageReads: make(map[common.Hash]struct{}),
|
StorageReads: make(map[common.Hash]struct{}),
|
||||||
BalanceChanges: make(map[uint16]*uint256.Int),
|
BalanceChanges: make(map[uint16]*uint256.Int),
|
||||||
NonceChanges: make(map[uint16]uint64),
|
NonceChanges: make(map[uint16]uint64),
|
||||||
|
CodeChange: make(map[uint16][]byte),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
@ -120,10 +114,8 @@ func (b *ConstructionBlockAccessList) CodeChange(address common.Address, txIndex
|
||||||
if _, ok := b.Accounts[address]; !ok {
|
if _, ok := b.Accounts[address]; !ok {
|
||||||
b.Accounts[address] = NewConstructionAccountAccess()
|
b.Accounts[address] = NewConstructionAccountAccess()
|
||||||
}
|
}
|
||||||
b.Accounts[address].CodeChange = &CodeChange{
|
// TODO(rjl493456442) is it essential to deep-copy the code?
|
||||||
TxIndex: txIndex,
|
b.Accounts[address].CodeChange[txIndex] = bytes.Clone(code)
|
||||||
Code: bytes.Clone(code),
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// NonceChange records tx post-state nonce of any contract-like accounts whose
|
// NonceChange records tx post-state nonce of any contract-like accounts whose
|
||||||
|
|
@ -170,12 +162,11 @@ func (b *ConstructionBlockAccessList) Copy() *ConstructionBlockAccessList {
|
||||||
aaCopy.BalanceChanges = balances
|
aaCopy.BalanceChanges = balances
|
||||||
aaCopy.NonceChanges = maps.Clone(aa.NonceChanges)
|
aaCopy.NonceChanges = maps.Clone(aa.NonceChanges)
|
||||||
|
|
||||||
if aa.CodeChange != nil {
|
codes := make(map[uint16][]byte, len(aa.CodeChange))
|
||||||
aaCopy.CodeChange = &CodeChange{
|
for index, code := range aa.CodeChange {
|
||||||
TxIndex: aa.CodeChange.TxIndex,
|
codes[index] = bytes.Clone(code)
|
||||||
Code: bytes.Clone(aa.CodeChange.Code),
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
aaCopy.CodeChange = codes
|
||||||
res.Accounts[addr] = &aaCopy
|
res.Accounts[addr] = &aaCopy
|
||||||
}
|
}
|
||||||
return &res
|
return &res
|
||||||
|
|
|
||||||
|
|
@ -119,6 +119,13 @@ func (e *encodingSlotWrites) validate() error {
|
||||||
return errors.New("storage write tx indices not in order")
|
return errors.New("storage write tx indices not in order")
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// encodingCodeChange contains the runtime bytecode deployed at an address
|
||||||
|
// and the transaction index where the deployment took place.
|
||||||
|
type encodingCodeChange struct {
|
||||||
|
TxIndex uint16 `ssz-size:"2"`
|
||||||
|
Code []byte `ssz-max:"300000"` // TODO(rjl493456442) shall we put the limit here? The limit will be increased gradually
|
||||||
|
}
|
||||||
|
|
||||||
// AccountAccess is the encoding format of ConstructionAccountAccess.
|
// AccountAccess is the encoding format of ConstructionAccountAccess.
|
||||||
type AccountAccess struct {
|
type AccountAccess struct {
|
||||||
Address [20]byte `ssz-size:"20"` // 20-byte Ethereum address
|
Address [20]byte `ssz-size:"20"` // 20-byte Ethereum address
|
||||||
|
|
@ -126,7 +133,7 @@ type AccountAccess struct {
|
||||||
StorageReads [][32]byte `ssz-max:"300000"` // Read-only storage keys
|
StorageReads [][32]byte `ssz-max:"300000"` // Read-only storage keys
|
||||||
BalanceChanges []encodingBalanceChange `ssz-max:"300000"` // Balance changes ([tx_index -> post_balance])
|
BalanceChanges []encodingBalanceChange `ssz-max:"300000"` // Balance changes ([tx_index -> post_balance])
|
||||||
NonceChanges []encodingAccountNonce `ssz-max:"300000"` // Nonce changes ([tx_index -> new_nonce])
|
NonceChanges []encodingAccountNonce `ssz-max:"300000"` // Nonce changes ([tx_index -> new_nonce])
|
||||||
Code []CodeChange `ssz-max:"1"` // Code changes ([tx_index -> new_code])
|
CodeChanges []encodingCodeChange `ssz-max:"300000"` // Code changes ([tx_index -> new_code])
|
||||||
}
|
}
|
||||||
|
|
||||||
// validate converts the account accesses out of encoding format.
|
// validate converts the account accesses out of encoding format.
|
||||||
|
|
@ -166,9 +173,16 @@ func (e *AccountAccess) validate() error {
|
||||||
return errors.New("nonce changes not in ascending order by tx index")
|
return errors.New("nonce changes not in ascending order by tx index")
|
||||||
}
|
}
|
||||||
|
|
||||||
// Convert code change
|
// Check the code changes are sorted in order
|
||||||
if len(e.Code) == 1 {
|
if !slices.IsSortedFunc(e.CodeChanges, func(a, b encodingCodeChange) int {
|
||||||
if len(e.Code[0].Code) > params.MaxCodeSize {
|
return cmp.Compare[uint16](a.TxIndex, b.TxIndex)
|
||||||
|
}) {
|
||||||
|
return errors.New("code changes not in ascending order by tx index")
|
||||||
|
}
|
||||||
|
for _, change := range e.CodeChanges {
|
||||||
|
// TODO(rjl493456442): This check should be fork-aware, since the limit may
|
||||||
|
// differ across forks.
|
||||||
|
if len(change.Code) > params.MaxCodeSize {
|
||||||
return errors.New("code change contained oversized code")
|
return errors.New("code change contained oversized code")
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
@ -182,6 +196,8 @@ func (e *AccountAccess) Copy() AccountAccess {
|
||||||
StorageReads: slices.Clone(e.StorageReads),
|
StorageReads: slices.Clone(e.StorageReads),
|
||||||
BalanceChanges: slices.Clone(e.BalanceChanges),
|
BalanceChanges: slices.Clone(e.BalanceChanges),
|
||||||
NonceChanges: slices.Clone(e.NonceChanges),
|
NonceChanges: slices.Clone(e.NonceChanges),
|
||||||
|
StorageWrites: make([]encodingSlotWrites, 0, len(e.StorageWrites)),
|
||||||
|
CodeChanges: make([]encodingCodeChange, 0, len(e.CodeChanges)),
|
||||||
}
|
}
|
||||||
for _, storageWrite := range e.StorageWrites {
|
for _, storageWrite := range e.StorageWrites {
|
||||||
res.StorageWrites = append(res.StorageWrites, encodingSlotWrites{
|
res.StorageWrites = append(res.StorageWrites, encodingSlotWrites{
|
||||||
|
|
@ -189,13 +205,11 @@ func (e *AccountAccess) Copy() AccountAccess {
|
||||||
Accesses: slices.Clone(storageWrite.Accesses),
|
Accesses: slices.Clone(storageWrite.Accesses),
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
if len(e.Code) == 1 {
|
for _, codeChange := range e.CodeChanges {
|
||||||
res.Code = []CodeChange{
|
res.CodeChanges = append(res.CodeChanges, encodingCodeChange{
|
||||||
{
|
TxIndex: codeChange.TxIndex,
|
||||||
e.Code[0].TxIndex,
|
Code: bytes.Clone(codeChange.Code),
|
||||||
bytes.Clone(e.Code[0].Code),
|
})
|
||||||
},
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
return res
|
return res
|
||||||
}
|
}
|
||||||
|
|
@ -212,11 +226,11 @@ var _ rlp.Encoder = &ConstructionBlockAccessList{}
|
||||||
func (a *ConstructionAccountAccess) toEncodingObj(addr common.Address) AccountAccess {
|
func (a *ConstructionAccountAccess) toEncodingObj(addr common.Address) AccountAccess {
|
||||||
res := AccountAccess{
|
res := AccountAccess{
|
||||||
Address: addr,
|
Address: addr,
|
||||||
StorageWrites: make([]encodingSlotWrites, 0),
|
StorageWrites: make([]encodingSlotWrites, 0, len(a.StorageWrites)),
|
||||||
StorageReads: make([][32]byte, 0),
|
StorageReads: make([][32]byte, 0, len(a.StorageReads)),
|
||||||
BalanceChanges: make([]encodingBalanceChange, 0),
|
BalanceChanges: make([]encodingBalanceChange, 0, len(a.BalanceChanges)),
|
||||||
NonceChanges: make([]encodingAccountNonce, 0),
|
NonceChanges: make([]encodingAccountNonce, 0, len(a.NonceChanges)),
|
||||||
Code: nil,
|
CodeChanges: make([]encodingCodeChange, 0, len(a.CodeChange)),
|
||||||
}
|
}
|
||||||
|
|
||||||
// Convert write slots
|
// Convert write slots
|
||||||
|
|
@ -268,13 +282,13 @@ func (a *ConstructionAccountAccess) toEncodingObj(addr common.Address) AccountAc
|
||||||
}
|
}
|
||||||
|
|
||||||
// Convert code change
|
// Convert code change
|
||||||
if a.CodeChange != nil {
|
codeIndices := slices.Collect(maps.Keys(a.CodeChange))
|
||||||
res.Code = []CodeChange{
|
slices.SortFunc(codeIndices, cmp.Compare[uint16])
|
||||||
{
|
for _, idx := range codeIndices {
|
||||||
a.CodeChange.TxIndex,
|
res.CodeChanges = append(res.CodeChanges, encodingCodeChange{
|
||||||
bytes.Clone(a.CodeChange.Code),
|
TxIndex: idx,
|
||||||
},
|
Code: a.CodeChange[idx],
|
||||||
}
|
})
|
||||||
}
|
}
|
||||||
return res
|
return res
|
||||||
}
|
}
|
||||||
|
|
@ -327,9 +341,9 @@ func (e *BlockAccessList) PrettyPrint() string {
|
||||||
printWithIndent(2, fmt.Sprintf("%d: %d", change.TxIdx, change.Nonce))
|
printWithIndent(2, fmt.Sprintf("%d: %d", change.TxIdx, change.Nonce))
|
||||||
}
|
}
|
||||||
|
|
||||||
if len(accountDiff.Code) > 0 {
|
printWithIndent(1, "code changes:")
|
||||||
printWithIndent(1, "code:")
|
for _, change := range accountDiff.CodeChanges {
|
||||||
printWithIndent(2, fmt.Sprintf("%d: %x", accountDiff.Code[0].TxIndex, accountDiff.Code[0].Code))
|
printWithIndent(2, fmt.Sprintf("%d: %x", change.TxIndex, change.Code))
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
return res.String()
|
return res.String()
|
||||||
|
|
|
||||||
|
|
@ -49,7 +49,7 @@ func (obj *BlockAccessList) EncodeRLP(_w io.Writer) error {
|
||||||
}
|
}
|
||||||
w.ListEnd(_tmp15)
|
w.ListEnd(_tmp15)
|
||||||
_tmp18 := w.List()
|
_tmp18 := w.List()
|
||||||
for _, _tmp19 := range _tmp2.Code {
|
for _, _tmp19 := range _tmp2.CodeChanges {
|
||||||
_tmp20 := w.List()
|
_tmp20 := w.List()
|
||||||
w.WriteUint64(uint64(_tmp19.TxIndex))
|
w.WriteUint64(uint64(_tmp19.TxIndex))
|
||||||
w.WriteBytes(_tmp19.Code)
|
w.WriteBytes(_tmp19.Code)
|
||||||
|
|
@ -228,13 +228,13 @@ func (obj *BlockAccessList) DecodeRLP(dec *rlp.Stream) error {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
_tmp2.NonceChanges = _tmp17
|
_tmp2.NonceChanges = _tmp17
|
||||||
// Code:
|
// CodeChanges:
|
||||||
var _tmp21 []CodeChange
|
var _tmp21 []encodingCodeChange
|
||||||
if _, err := dec.List(); err != nil {
|
if _, err := dec.List(); err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
for dec.MoreDataInList() {
|
for dec.MoreDataInList() {
|
||||||
var _tmp22 CodeChange
|
var _tmp22 encodingCodeChange
|
||||||
{
|
{
|
||||||
if _, err := dec.List(); err != nil {
|
if _, err := dec.List(); err != nil {
|
||||||
return err
|
return err
|
||||||
|
|
@ -260,7 +260,7 @@ func (obj *BlockAccessList) DecodeRLP(dec *rlp.Stream) error {
|
||||||
if err := dec.ListEnd(); err != nil {
|
if err := dec.ListEnd(); err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
_tmp2.Code = _tmp21
|
_tmp2.CodeChanges = _tmp21
|
||||||
if err := dec.ListEnd(); err != nil {
|
if err := dec.ListEnd(); err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
|
|
||||||
|
|
@ -60,9 +60,8 @@ func makeTestConstructionBAL() *ConstructionBlockAccessList {
|
||||||
1: 2,
|
1: 2,
|
||||||
2: 6,
|
2: 6,
|
||||||
},
|
},
|
||||||
CodeChange: &CodeChange{
|
CodeChange: map[uint16][]byte{
|
||||||
TxIndex: 0,
|
0: common.Hex2Bytes("deadbeef"),
|
||||||
Code: common.Hex2Bytes("deadbeef"),
|
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
common.BytesToAddress([]byte{0xff, 0xff, 0xff}): {
|
common.BytesToAddress([]byte{0xff, 0xff, 0xff}): {
|
||||||
|
|
@ -85,6 +84,9 @@ func makeTestConstructionBAL() *ConstructionBlockAccessList {
|
||||||
NonceChanges: map[uint16]uint64{
|
NonceChanges: map[uint16]uint64{
|
||||||
1: 2,
|
1: 2,
|
||||||
},
|
},
|
||||||
|
CodeChange: map[uint16][]byte{
|
||||||
|
0: common.Hex2Bytes("deadbeef"),
|
||||||
|
},
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
}
|
}
|
||||||
|
|
@ -179,7 +181,7 @@ func makeTestAccountAccess(sort bool) AccountAccess {
|
||||||
StorageReads: storageReads,
|
StorageReads: storageReads,
|
||||||
BalanceChanges: balances,
|
BalanceChanges: balances,
|
||||||
NonceChanges: nonces,
|
NonceChanges: nonces,
|
||||||
Code: []CodeChange{
|
CodeChanges: []encodingCodeChange{
|
||||||
{
|
{
|
||||||
TxIndex: 100,
|
TxIndex: 100,
|
||||||
Code: testrand.Bytes(256),
|
Code: testrand.Bytes(256),
|
||||||
|
|
|
||||||
|
|
@ -98,6 +98,9 @@ type Header struct {
|
||||||
|
|
||||||
// RequestsHash was added by EIP-7685 and is ignored in legacy headers.
|
// RequestsHash was added by EIP-7685 and is ignored in legacy headers.
|
||||||
RequestsHash *common.Hash `json:"requestsHash" rlp:"optional"`
|
RequestsHash *common.Hash `json:"requestsHash" rlp:"optional"`
|
||||||
|
|
||||||
|
// SlotNumber was added by EIP-7843 and is ignored in legacy headers.
|
||||||
|
SlotNumber *uint64 `json:"slotNumber" rlp:"optional"`
|
||||||
}
|
}
|
||||||
|
|
||||||
// field type overrides for gencodec
|
// field type overrides for gencodec
|
||||||
|
|
@ -112,6 +115,7 @@ type headerMarshaling struct {
|
||||||
Hash common.Hash `json:"hash"` // adds call to Hash() in MarshalJSON
|
Hash common.Hash `json:"hash"` // adds call to Hash() in MarshalJSON
|
||||||
BlobGasUsed *hexutil.Uint64
|
BlobGasUsed *hexutil.Uint64
|
||||||
ExcessBlobGas *hexutil.Uint64
|
ExcessBlobGas *hexutil.Uint64
|
||||||
|
SlotNumber *hexutil.Uint64
|
||||||
}
|
}
|
||||||
|
|
||||||
// Hash returns the block hash of the header, which is simply the keccak256 hash of its
|
// Hash returns the block hash of the header, which is simply the keccak256 hash of its
|
||||||
|
|
@ -316,6 +320,10 @@ func CopyHeader(h *Header) *Header {
|
||||||
cpy.RequestsHash = new(common.Hash)
|
cpy.RequestsHash = new(common.Hash)
|
||||||
*cpy.RequestsHash = *h.RequestsHash
|
*cpy.RequestsHash = *h.RequestsHash
|
||||||
}
|
}
|
||||||
|
if h.SlotNumber != nil {
|
||||||
|
cpy.SlotNumber = new(uint64)
|
||||||
|
*cpy.SlotNumber = *h.SlotNumber
|
||||||
|
}
|
||||||
return &cpy
|
return &cpy
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
@ -416,6 +424,15 @@ func (b *Block) BlobGasUsed() *uint64 {
|
||||||
return blobGasUsed
|
return blobGasUsed
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func (b *Block) SlotNumber() *uint64 {
|
||||||
|
var slotNum *uint64
|
||||||
|
if b.header.SlotNumber != nil {
|
||||||
|
slotNum = new(uint64)
|
||||||
|
*slotNum = *b.header.SlotNumber
|
||||||
|
}
|
||||||
|
return slotNum
|
||||||
|
}
|
||||||
|
|
||||||
// Size returns the true RLP encoded storage size of the block, either by encoding
|
// Size returns the true RLP encoded storage size of the block, either by encoding
|
||||||
// and returning it, or returning a previously cached value.
|
// and returning it, or returning a previously cached value.
|
||||||
func (b *Block) Size() uint64 {
|
func (b *Block) Size() uint64 {
|
||||||
|
|
|
||||||
|
|
@ -37,6 +37,7 @@ func (h Header) MarshalJSON() ([]byte, error) {
|
||||||
ExcessBlobGas *hexutil.Uint64 `json:"excessBlobGas" rlp:"optional"`
|
ExcessBlobGas *hexutil.Uint64 `json:"excessBlobGas" rlp:"optional"`
|
||||||
ParentBeaconRoot *common.Hash `json:"parentBeaconBlockRoot" rlp:"optional"`
|
ParentBeaconRoot *common.Hash `json:"parentBeaconBlockRoot" rlp:"optional"`
|
||||||
RequestsHash *common.Hash `json:"requestsHash" rlp:"optional"`
|
RequestsHash *common.Hash `json:"requestsHash" rlp:"optional"`
|
||||||
|
SlotNumber *hexutil.Uint64 `json:"slotNumber" rlp:"optional"`
|
||||||
Hash common.Hash `json:"hash"`
|
Hash common.Hash `json:"hash"`
|
||||||
}
|
}
|
||||||
var enc Header
|
var enc Header
|
||||||
|
|
@ -61,6 +62,7 @@ func (h Header) MarshalJSON() ([]byte, error) {
|
||||||
enc.ExcessBlobGas = (*hexutil.Uint64)(h.ExcessBlobGas)
|
enc.ExcessBlobGas = (*hexutil.Uint64)(h.ExcessBlobGas)
|
||||||
enc.ParentBeaconRoot = h.ParentBeaconRoot
|
enc.ParentBeaconRoot = h.ParentBeaconRoot
|
||||||
enc.RequestsHash = h.RequestsHash
|
enc.RequestsHash = h.RequestsHash
|
||||||
|
enc.SlotNumber = (*hexutil.Uint64)(h.SlotNumber)
|
||||||
enc.Hash = h.Hash()
|
enc.Hash = h.Hash()
|
||||||
return json.Marshal(&enc)
|
return json.Marshal(&enc)
|
||||||
}
|
}
|
||||||
|
|
@ -89,6 +91,7 @@ func (h *Header) UnmarshalJSON(input []byte) error {
|
||||||
ExcessBlobGas *hexutil.Uint64 `json:"excessBlobGas" rlp:"optional"`
|
ExcessBlobGas *hexutil.Uint64 `json:"excessBlobGas" rlp:"optional"`
|
||||||
ParentBeaconRoot *common.Hash `json:"parentBeaconBlockRoot" rlp:"optional"`
|
ParentBeaconRoot *common.Hash `json:"parentBeaconBlockRoot" rlp:"optional"`
|
||||||
RequestsHash *common.Hash `json:"requestsHash" rlp:"optional"`
|
RequestsHash *common.Hash `json:"requestsHash" rlp:"optional"`
|
||||||
|
SlotNumber *hexutil.Uint64 `json:"slotNumber" rlp:"optional"`
|
||||||
}
|
}
|
||||||
var dec Header
|
var dec Header
|
||||||
if err := json.Unmarshal(input, &dec); err != nil {
|
if err := json.Unmarshal(input, &dec); err != nil {
|
||||||
|
|
@ -169,5 +172,8 @@ func (h *Header) UnmarshalJSON(input []byte) error {
|
||||||
if dec.RequestsHash != nil {
|
if dec.RequestsHash != nil {
|
||||||
h.RequestsHash = dec.RequestsHash
|
h.RequestsHash = dec.RequestsHash
|
||||||
}
|
}
|
||||||
|
if dec.SlotNumber != nil {
|
||||||
|
h.SlotNumber = (*uint64)(dec.SlotNumber)
|
||||||
|
}
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
|
||||||
|
|
@ -43,7 +43,8 @@ func (obj *Header) EncodeRLP(_w io.Writer) error {
|
||||||
_tmp4 := obj.ExcessBlobGas != nil
|
_tmp4 := obj.ExcessBlobGas != nil
|
||||||
_tmp5 := obj.ParentBeaconRoot != nil
|
_tmp5 := obj.ParentBeaconRoot != nil
|
||||||
_tmp6 := obj.RequestsHash != nil
|
_tmp6 := obj.RequestsHash != nil
|
||||||
if _tmp1 || _tmp2 || _tmp3 || _tmp4 || _tmp5 || _tmp6 {
|
_tmp7 := obj.SlotNumber != nil
|
||||||
|
if _tmp1 || _tmp2 || _tmp3 || _tmp4 || _tmp5 || _tmp6 || _tmp7 {
|
||||||
if obj.BaseFee == nil {
|
if obj.BaseFee == nil {
|
||||||
w.Write(rlp.EmptyString)
|
w.Write(rlp.EmptyString)
|
||||||
} else {
|
} else {
|
||||||
|
|
@ -53,41 +54,48 @@ func (obj *Header) EncodeRLP(_w io.Writer) error {
|
||||||
w.WriteBigInt(obj.BaseFee)
|
w.WriteBigInt(obj.BaseFee)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
if _tmp2 || _tmp3 || _tmp4 || _tmp5 || _tmp6 {
|
if _tmp2 || _tmp3 || _tmp4 || _tmp5 || _tmp6 || _tmp7 {
|
||||||
if obj.WithdrawalsHash == nil {
|
if obj.WithdrawalsHash == nil {
|
||||||
w.Write([]byte{0x80})
|
w.Write([]byte{0x80})
|
||||||
} else {
|
} else {
|
||||||
w.WriteBytes(obj.WithdrawalsHash[:])
|
w.WriteBytes(obj.WithdrawalsHash[:])
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
if _tmp3 || _tmp4 || _tmp5 || _tmp6 {
|
if _tmp3 || _tmp4 || _tmp5 || _tmp6 || _tmp7 {
|
||||||
if obj.BlobGasUsed == nil {
|
if obj.BlobGasUsed == nil {
|
||||||
w.Write([]byte{0x80})
|
w.Write([]byte{0x80})
|
||||||
} else {
|
} else {
|
||||||
w.WriteUint64((*obj.BlobGasUsed))
|
w.WriteUint64((*obj.BlobGasUsed))
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
if _tmp4 || _tmp5 || _tmp6 {
|
if _tmp4 || _tmp5 || _tmp6 || _tmp7 {
|
||||||
if obj.ExcessBlobGas == nil {
|
if obj.ExcessBlobGas == nil {
|
||||||
w.Write([]byte{0x80})
|
w.Write([]byte{0x80})
|
||||||
} else {
|
} else {
|
||||||
w.WriteUint64((*obj.ExcessBlobGas))
|
w.WriteUint64((*obj.ExcessBlobGas))
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
if _tmp5 || _tmp6 {
|
if _tmp5 || _tmp6 || _tmp7 {
|
||||||
if obj.ParentBeaconRoot == nil {
|
if obj.ParentBeaconRoot == nil {
|
||||||
w.Write([]byte{0x80})
|
w.Write([]byte{0x80})
|
||||||
} else {
|
} else {
|
||||||
w.WriteBytes(obj.ParentBeaconRoot[:])
|
w.WriteBytes(obj.ParentBeaconRoot[:])
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
if _tmp6 {
|
if _tmp6 || _tmp7 {
|
||||||
if obj.RequestsHash == nil {
|
if obj.RequestsHash == nil {
|
||||||
w.Write([]byte{0x80})
|
w.Write([]byte{0x80})
|
||||||
} else {
|
} else {
|
||||||
w.WriteBytes(obj.RequestsHash[:])
|
w.WriteBytes(obj.RequestsHash[:])
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
if _tmp7 {
|
||||||
|
if obj.SlotNumber == nil {
|
||||||
|
w.Write([]byte{0x80})
|
||||||
|
} else {
|
||||||
|
w.WriteUint64((*obj.SlotNumber))
|
||||||
|
}
|
||||||
|
}
|
||||||
w.ListEnd(_tmp0)
|
w.ListEnd(_tmp0)
|
||||||
return w.Flush()
|
return w.Flush()
|
||||||
}
|
}
|
||||||
|
|
|
||||||
|
|
@ -396,14 +396,37 @@ func (tx *Transaction) calcEffectiveGasTip(dst *uint256.Int, baseFee *uint256.In
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// EffectiveGasTipValue returns the effective gasTip value for the given base fee,
|
||||||
|
// even if it would be negative. This can be used for sorting purposes.
|
||||||
|
func (tx *Transaction) EffectiveGasTipValue(baseFee *big.Int) *big.Int {
|
||||||
|
// min(gasTipCap, gasFeeCap - baseFee)
|
||||||
|
dst := new(big.Int)
|
||||||
|
if baseFee == nil {
|
||||||
|
dst.Set(tx.inner.gasTipCap())
|
||||||
|
return dst
|
||||||
|
}
|
||||||
|
|
||||||
|
dst.Sub(tx.inner.gasFeeCap(), baseFee) // gasFeeCap - baseFee
|
||||||
|
gasTipCap := tx.inner.gasTipCap()
|
||||||
|
if gasTipCap.Cmp(dst) < 0 { // gasTipCap < (gasFeeCap - baseFee)
|
||||||
|
dst.Set(gasTipCap)
|
||||||
|
}
|
||||||
|
return dst
|
||||||
|
}
|
||||||
|
|
||||||
func (tx *Transaction) EffectiveGasTipCmp(other *Transaction, baseFee *uint256.Int) int {
|
func (tx *Transaction) EffectiveGasTipCmp(other *Transaction, baseFee *uint256.Int) int {
|
||||||
if baseFee == nil {
|
if baseFee == nil {
|
||||||
return tx.GasTipCapCmp(other)
|
return tx.GasTipCapCmp(other)
|
||||||
}
|
}
|
||||||
// Use more efficient internal method.
|
// Use more efficient internal method.
|
||||||
txTip, otherTip := new(uint256.Int), new(uint256.Int)
|
txTip, otherTip := new(uint256.Int), new(uint256.Int)
|
||||||
tx.calcEffectiveGasTip(txTip, baseFee)
|
err1 := tx.calcEffectiveGasTip(txTip, baseFee)
|
||||||
other.calcEffectiveGasTip(otherTip, baseFee)
|
err2 := other.calcEffectiveGasTip(otherTip, baseFee)
|
||||||
|
if err1 != nil || err2 != nil {
|
||||||
|
// fall back to big int comparison in case of error
|
||||||
|
base := baseFee.ToBig()
|
||||||
|
return tx.EffectiveGasTipValue(base).Cmp(other.EffectiveGasTipValue(base))
|
||||||
|
}
|
||||||
return txTip.Cmp(otherTip)
|
return txTip.Cmp(otherTip)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
||||||
|
|
@ -262,7 +262,7 @@ func ActivePrecompiles(rules params.Rules) []common.Address {
|
||||||
// - the returned bytes,
|
// - the returned bytes,
|
||||||
// - the _remaining_ gas,
|
// - the _remaining_ gas,
|
||||||
// - any error that occurred
|
// - any error that occurred
|
||||||
func RunPrecompiledContract(p PrecompiledContract, input []byte, suppliedGas uint64, logger *tracing.Hooks) (ret []byte, remainingGas uint64, err error) {
|
func RunPrecompiledContract(stateDB StateDB, p PrecompiledContract, address common.Address, input []byte, suppliedGas uint64, logger *tracing.Hooks) (ret []byte, remainingGas uint64, err error) {
|
||||||
gasCost := p.RequiredGas(input)
|
gasCost := p.RequiredGas(input)
|
||||||
if suppliedGas < gasCost {
|
if suppliedGas < gasCost {
|
||||||
return nil, 0, ErrOutOfGas
|
return nil, 0, ErrOutOfGas
|
||||||
|
|
@ -271,6 +271,12 @@ func RunPrecompiledContract(p PrecompiledContract, input []byte, suppliedGas uin
|
||||||
logger.OnGasChange(suppliedGas, suppliedGas-gasCost, tracing.GasChangeCallPrecompiledContract)
|
logger.OnGasChange(suppliedGas, suppliedGas-gasCost, tracing.GasChangeCallPrecompiledContract)
|
||||||
}
|
}
|
||||||
suppliedGas -= gasCost
|
suppliedGas -= gasCost
|
||||||
|
|
||||||
|
// Touch the precompile for block-level accessList recording once Amsterdam
|
||||||
|
// fork is activated.
|
||||||
|
if stateDB != nil {
|
||||||
|
stateDB.Exist(address)
|
||||||
|
}
|
||||||
output, err := p.Run(input)
|
output, err := p.Run(input)
|
||||||
return output, suppliedGas, err
|
return output, suppliedGas, err
|
||||||
}
|
}
|
||||||
|
|
|
||||||
|
|
@ -36,7 +36,7 @@ func FuzzPrecompiledContracts(f *testing.F) {
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
inWant := string(input)
|
inWant := string(input)
|
||||||
RunPrecompiledContract(p, input, gas, nil)
|
RunPrecompiledContract(nil, p, a, input, gas, nil)
|
||||||
if inHave := string(input); inWant != inHave {
|
if inHave := string(input); inWant != inHave {
|
||||||
t.Errorf("Precompiled %v modified input data", a)
|
t.Errorf("Precompiled %v modified input data", a)
|
||||||
}
|
}
|
||||||
|
|
|
||||||
|
|
@ -99,7 +99,7 @@ func testPrecompiled(addr string, test precompiledTest, t *testing.T) {
|
||||||
in := common.Hex2Bytes(test.Input)
|
in := common.Hex2Bytes(test.Input)
|
||||||
gas := p.RequiredGas(in)
|
gas := p.RequiredGas(in)
|
||||||
t.Run(fmt.Sprintf("%s-Gas=%d", test.Name, gas), func(t *testing.T) {
|
t.Run(fmt.Sprintf("%s-Gas=%d", test.Name, gas), func(t *testing.T) {
|
||||||
if res, _, err := RunPrecompiledContract(p, in, gas, nil); err != nil {
|
if res, _, err := RunPrecompiledContract(nil, p, common.HexToAddress(addr), in, gas, nil); err != nil {
|
||||||
t.Error(err)
|
t.Error(err)
|
||||||
} else if common.Bytes2Hex(res) != test.Expected {
|
} else if common.Bytes2Hex(res) != test.Expected {
|
||||||
t.Errorf("Expected %v, got %v", test.Expected, common.Bytes2Hex(res))
|
t.Errorf("Expected %v, got %v", test.Expected, common.Bytes2Hex(res))
|
||||||
|
|
@ -121,7 +121,7 @@ func testPrecompiledOOG(addr string, test precompiledTest, t *testing.T) {
|
||||||
gas := test.Gas - 1
|
gas := test.Gas - 1
|
||||||
|
|
||||||
t.Run(fmt.Sprintf("%s-Gas=%d", test.Name, gas), func(t *testing.T) {
|
t.Run(fmt.Sprintf("%s-Gas=%d", test.Name, gas), func(t *testing.T) {
|
||||||
_, _, err := RunPrecompiledContract(p, in, gas, nil)
|
_, _, err := RunPrecompiledContract(nil, p, common.HexToAddress(addr), in, gas, nil)
|
||||||
if err.Error() != "out of gas" {
|
if err.Error() != "out of gas" {
|
||||||
t.Errorf("Expected error [out of gas], got [%v]", err)
|
t.Errorf("Expected error [out of gas], got [%v]", err)
|
||||||
}
|
}
|
||||||
|
|
@ -138,7 +138,7 @@ func testPrecompiledFailure(addr string, test precompiledFailureTest, t *testing
|
||||||
in := common.Hex2Bytes(test.Input)
|
in := common.Hex2Bytes(test.Input)
|
||||||
gas := p.RequiredGas(in)
|
gas := p.RequiredGas(in)
|
||||||
t.Run(test.Name, func(t *testing.T) {
|
t.Run(test.Name, func(t *testing.T) {
|
||||||
_, _, err := RunPrecompiledContract(p, in, gas, nil)
|
_, _, err := RunPrecompiledContract(nil, p, common.HexToAddress(addr), in, gas, nil)
|
||||||
if err.Error() != test.ExpectedError {
|
if err.Error() != test.ExpectedError {
|
||||||
t.Errorf("Expected error [%v], got [%v]", test.ExpectedError, err)
|
t.Errorf("Expected error [%v], got [%v]", test.ExpectedError, err)
|
||||||
}
|
}
|
||||||
|
|
@ -169,7 +169,7 @@ func benchmarkPrecompiled(addr string, test precompiledTest, bench *testing.B) {
|
||||||
start := time.Now()
|
start := time.Now()
|
||||||
for bench.Loop() {
|
for bench.Loop() {
|
||||||
copy(data, in)
|
copy(data, in)
|
||||||
res, _, err = RunPrecompiledContract(p, data, reqGas, nil)
|
res, _, err = RunPrecompiledContract(nil, p, common.HexToAddress(addr), data, reqGas, nil)
|
||||||
}
|
}
|
||||||
elapsed := uint64(time.Since(start))
|
elapsed := uint64(time.Since(start))
|
||||||
if elapsed < 1 {
|
if elapsed < 1 {
|
||||||
|
|
|
||||||
|
|
@ -43,6 +43,7 @@ var activators = map[int]func(*JumpTable){
|
||||||
7702: enable7702,
|
7702: enable7702,
|
||||||
7939: enable7939,
|
7939: enable7939,
|
||||||
8024: enable8024,
|
8024: enable8024,
|
||||||
|
7843: enable7843,
|
||||||
}
|
}
|
||||||
|
|
||||||
// EnableEIP enables the given EIP on the config.
|
// EnableEIP enables the given EIP on the config.
|
||||||
|
|
@ -579,3 +580,19 @@ func enable7702(jt *JumpTable) {
|
||||||
jt[STATICCALL].dynamicGas = gasStaticCallEIP7702
|
jt[STATICCALL].dynamicGas = gasStaticCallEIP7702
|
||||||
jt[DELEGATECALL].dynamicGas = gasDelegateCallEIP7702
|
jt[DELEGATECALL].dynamicGas = gasDelegateCallEIP7702
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// opSlotNum enables the SLOTNUM opcode
|
||||||
|
func opSlotNum(pc *uint64, evm *EVM, scope *ScopeContext) ([]byte, error) {
|
||||||
|
scope.Stack.push(uint256.NewInt(evm.Context.SlotNum))
|
||||||
|
return nil, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// enable7843 enables the SLOTNUM opcode as specified in EIP-7843.
|
||||||
|
func enable7843(jt *JumpTable) {
|
||||||
|
jt[SLOTNUM] = &operation{
|
||||||
|
execute: opSlotNum,
|
||||||
|
constantGas: GasQuickStep,
|
||||||
|
minStack: minStack(0, 1),
|
||||||
|
maxStack: maxStack(0, 1),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
|
||||||
|
|
@ -66,6 +66,7 @@ type BlockContext struct {
|
||||||
BaseFee *big.Int // Provides information for BASEFEE (0 if vm runs with NoBaseFee flag and 0 gas price)
|
BaseFee *big.Int // Provides information for BASEFEE (0 if vm runs with NoBaseFee flag and 0 gas price)
|
||||||
BlobBaseFee *big.Int // Provides information for BLOBBASEFEE (0 if vm runs with NoBaseFee flag and 0 blob gas price)
|
BlobBaseFee *big.Int // Provides information for BLOBBASEFEE (0 if vm runs with NoBaseFee flag and 0 blob gas price)
|
||||||
Random *common.Hash // Provides information for PREVRANDAO
|
Random *common.Hash // Provides information for PREVRANDAO
|
||||||
|
SlotNum uint64 // Provides information for SLOTNUM
|
||||||
}
|
}
|
||||||
|
|
||||||
// TxContext provides the EVM with information about a transaction.
|
// TxContext provides the EVM with information about a transaction.
|
||||||
|
|
@ -144,6 +145,8 @@ func NewEVM(blockCtx BlockContext, statedb StateDB, chainConfig *params.ChainCon
|
||||||
evm.precompiles = activePrecompiledContracts(evm.chainRules)
|
evm.precompiles = activePrecompiledContracts(evm.chainRules)
|
||||||
|
|
||||||
switch {
|
switch {
|
||||||
|
case evm.chainRules.IsAmsterdam:
|
||||||
|
evm.table = &amsterdamInstructionSet
|
||||||
case evm.chainRules.IsOsaka:
|
case evm.chainRules.IsOsaka:
|
||||||
evm.table = &osakaInstructionSet
|
evm.table = &osakaInstructionSet
|
||||||
case evm.chainRules.IsVerkle:
|
case evm.chainRules.IsVerkle:
|
||||||
|
|
@ -245,13 +248,14 @@ func (evm *EVM) Call(caller common.Address, addr common.Address, input []byte, g
|
||||||
if evm.depth > int(params.CallCreateDepth) {
|
if evm.depth > int(params.CallCreateDepth) {
|
||||||
return nil, gas, ErrDepth
|
return nil, gas, ErrDepth
|
||||||
}
|
}
|
||||||
// Fail if we're trying to transfer more than the available balance
|
syscall := isSystemCall(caller)
|
||||||
if !value.IsZero() && !evm.Context.CanTransfer(evm.StateDB, caller, value) {
|
|
||||||
|
// Fail if we're trying to transfer more than the available balance.
|
||||||
|
if !syscall && !value.IsZero() && !evm.Context.CanTransfer(evm.StateDB, caller, value) {
|
||||||
return nil, gas, ErrInsufficientBalance
|
return nil, gas, ErrInsufficientBalance
|
||||||
}
|
}
|
||||||
snapshot := evm.StateDB.Snapshot()
|
snapshot := evm.StateDB.Snapshot()
|
||||||
p, isPrecompile := evm.precompile(addr)
|
p, isPrecompile := evm.precompile(addr)
|
||||||
|
|
||||||
if !evm.StateDB.Exist(addr) {
|
if !evm.StateDB.Exist(addr) {
|
||||||
if !isPrecompile && evm.chainRules.IsEIP4762 && !isSystemCall(caller) {
|
if !isPrecompile && evm.chainRules.IsEIP4762 && !isSystemCall(caller) {
|
||||||
// Add proof of absence to witness
|
// Add proof of absence to witness
|
||||||
|
|
@ -275,10 +279,18 @@ func (evm *EVM) Call(caller common.Address, addr common.Address, input []byte, g
|
||||||
}
|
}
|
||||||
evm.StateDB.CreateAccount(addr)
|
evm.StateDB.CreateAccount(addr)
|
||||||
}
|
}
|
||||||
evm.Context.Transfer(evm.StateDB, caller, addr, value)
|
// Perform the value transfer only in non-syscall mode.
|
||||||
|
// Calling this is required even for zero-value transfers,
|
||||||
|
// to ensure the state clearing mechanism is applied.
|
||||||
|
if !syscall {
|
||||||
|
evm.Context.Transfer(evm.StateDB, caller, addr, value)
|
||||||
|
}
|
||||||
if isPrecompile {
|
if isPrecompile {
|
||||||
ret, gas, err = RunPrecompiledContract(p, input, gas, evm.Config.Tracer)
|
var stateDB StateDB
|
||||||
|
if evm.chainRules.IsAmsterdam {
|
||||||
|
stateDB = evm.StateDB
|
||||||
|
}
|
||||||
|
ret, gas, err = RunPrecompiledContract(stateDB, p, addr, input, gas, evm.Config.Tracer)
|
||||||
} else {
|
} else {
|
||||||
// Initialise a new contract and set the code that is to be used by the EVM.
|
// Initialise a new contract and set the code that is to be used by the EVM.
|
||||||
code := evm.resolveCode(addr)
|
code := evm.resolveCode(addr)
|
||||||
|
|
@ -302,7 +314,6 @@ func (evm *EVM) Call(caller common.Address, addr common.Address, input []byte, g
|
||||||
if evm.Config.Tracer != nil && evm.Config.Tracer.OnGasChange != nil {
|
if evm.Config.Tracer != nil && evm.Config.Tracer.OnGasChange != nil {
|
||||||
evm.Config.Tracer.OnGasChange(gas, 0, tracing.GasChangeCallFailedExecution)
|
evm.Config.Tracer.OnGasChange(gas, 0, tracing.GasChangeCallFailedExecution)
|
||||||
}
|
}
|
||||||
|
|
||||||
gas = 0
|
gas = 0
|
||||||
}
|
}
|
||||||
// TODO: consider clearing up unused snapshots:
|
// TODO: consider clearing up unused snapshots:
|
||||||
|
|
@ -342,7 +353,11 @@ func (evm *EVM) CallCode(caller common.Address, addr common.Address, input []byt
|
||||||
|
|
||||||
// It is allowed to call precompiles, even via delegatecall
|
// It is allowed to call precompiles, even via delegatecall
|
||||||
if p, isPrecompile := evm.precompile(addr); isPrecompile {
|
if p, isPrecompile := evm.precompile(addr); isPrecompile {
|
||||||
ret, gas, err = RunPrecompiledContract(p, input, gas, evm.Config.Tracer)
|
var stateDB StateDB
|
||||||
|
if evm.chainRules.IsAmsterdam {
|
||||||
|
stateDB = evm.StateDB
|
||||||
|
}
|
||||||
|
ret, gas, err = RunPrecompiledContract(stateDB, p, addr, input, gas, evm.Config.Tracer)
|
||||||
} else {
|
} else {
|
||||||
// Initialise a new contract and set the code that is to be used by the EVM.
|
// Initialise a new contract and set the code that is to be used by the EVM.
|
||||||
// The contract is a scoped environment for this execution context only.
|
// The contract is a scoped environment for this execution context only.
|
||||||
|
|
@ -385,7 +400,11 @@ func (evm *EVM) DelegateCall(originCaller common.Address, caller common.Address,
|
||||||
|
|
||||||
// It is allowed to call precompiles, even via delegatecall
|
// It is allowed to call precompiles, even via delegatecall
|
||||||
if p, isPrecompile := evm.precompile(addr); isPrecompile {
|
if p, isPrecompile := evm.precompile(addr); isPrecompile {
|
||||||
ret, gas, err = RunPrecompiledContract(p, input, gas, evm.Config.Tracer)
|
var stateDB StateDB
|
||||||
|
if evm.chainRules.IsAmsterdam {
|
||||||
|
stateDB = evm.StateDB
|
||||||
|
}
|
||||||
|
ret, gas, err = RunPrecompiledContract(stateDB, p, addr, input, gas, evm.Config.Tracer)
|
||||||
} else {
|
} else {
|
||||||
// Initialise a new contract and make initialise the delegate values
|
// Initialise a new contract and make initialise the delegate values
|
||||||
//
|
//
|
||||||
|
|
@ -437,7 +456,11 @@ func (evm *EVM) StaticCall(caller common.Address, addr common.Address, input []b
|
||||||
evm.StateDB.AddBalance(addr, new(uint256.Int), tracing.BalanceChangeTouchAccount)
|
evm.StateDB.AddBalance(addr, new(uint256.Int), tracing.BalanceChangeTouchAccount)
|
||||||
|
|
||||||
if p, isPrecompile := evm.precompile(addr); isPrecompile {
|
if p, isPrecompile := evm.precompile(addr); isPrecompile {
|
||||||
ret, gas, err = RunPrecompiledContract(p, input, gas, evm.Config.Tracer)
|
var stateDB StateDB
|
||||||
|
if evm.chainRules.IsAmsterdam {
|
||||||
|
stateDB = evm.StateDB
|
||||||
|
}
|
||||||
|
ret, gas, err = RunPrecompiledContract(stateDB, p, addr, input, gas, evm.Config.Tracer)
|
||||||
} else {
|
} else {
|
||||||
// Initialise a new contract and set the code that is to be used by the EVM.
|
// Initialise a new contract and set the code that is to be used by the EVM.
|
||||||
// The contract is a scoped environment for this execution context only.
|
// The contract is a scoped environment for this execution context only.
|
||||||
|
|
|
||||||
|
|
@ -946,24 +946,34 @@ func opSelfdestruct6780(pc *uint64, evm *EVM, scope *ScopeContext) ([]byte, erro
|
||||||
return nil, errStopToken
|
return nil, errStopToken
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// decodeSingle decodes the immediate operand of a backward-compatible DUPN or SWAPN instruction (EIP-8024)
|
||||||
|
// https://eips.ethereum.org/EIPS/eip-8024
|
||||||
func decodeSingle(x byte) int {
|
func decodeSingle(x byte) int {
|
||||||
if x <= 90 {
|
// Depths 1-16 are already covered by the legacy opcodes. The forbidden byte range [91, 127] removes
|
||||||
return int(x) + 17
|
// 37 values from the 256 possible immediates, leaving 219 usable values, so this encoding covers depths
|
||||||
}
|
// 17 through 235. The immediate is encoded as (x + 111) % 256, where 111 is chosen so that these values
|
||||||
return int(x) - 20
|
// avoid the forbidden range. Decoding is simply the modular inverse (i.e. 111+145=256).
|
||||||
|
return (int(x) + 145) % 256
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// decodePair decodes the immediate operand of a backward-compatible EXCHANGE
|
||||||
|
// instruction (EIP-8024) into stack indices (n, m) where 1 <= n < m
|
||||||
|
// and n + m <= 30. The forbidden byte range [82, 127] removes 46 values from
|
||||||
|
// the 256 possible immediates, leaving exactly 210 usable bytes.
|
||||||
|
// https://eips.ethereum.org/EIPS/eip-8024
|
||||||
func decodePair(x byte) (int, int) {
|
func decodePair(x byte) (int, int) {
|
||||||
var k int
|
// XOR with 143 remaps the forbidden bytes [82, 127] to an unused corner
|
||||||
if x <= 79 {
|
// of the 16x16 grid below.
|
||||||
k = int(x)
|
k := int(x ^ 143)
|
||||||
} else {
|
// Split into row q and column r of a 16x16 grid. The 210 valid pairs
|
||||||
k = int(x) - 48
|
// occupy two triangles within this grid.
|
||||||
}
|
|
||||||
q, r := k/16, k%16
|
q, r := k/16, k%16
|
||||||
|
// Upper triangle (q < r): pairs where m <= 16, encoded directly as
|
||||||
|
// (q+1, r+1).
|
||||||
if q < r {
|
if q < r {
|
||||||
return q + 1, r + 1
|
return q + 1, r + 1
|
||||||
}
|
}
|
||||||
|
// Lower triangle: pairs where m > 16, recovered as (r+1, 29-q).
|
||||||
return r + 1, 29 - q
|
return r + 1, 29 - q
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
@ -1034,8 +1044,8 @@ func opExchange(pc *uint64, evm *EVM, scope *ScopeContext) ([]byte, error) {
|
||||||
}
|
}
|
||||||
|
|
||||||
// This range is excluded both to preserve compatibility with existing opcodes
|
// This range is excluded both to preserve compatibility with existing opcodes
|
||||||
// and to keep decode_pair’s 16-aligned arithmetic mapping valid (0–79, 128–255).
|
// and to keep decode_pair’s 16-aligned arithmetic mapping valid (0–81, 128–255).
|
||||||
if x > 79 && x < 128 {
|
if x > 81 && x < 128 {
|
||||||
return nil, &ErrInvalidOpCode{opcode: OpCode(x)}
|
return nil, &ErrInvalidOpCode{opcode: OpCode(x)}
|
||||||
}
|
}
|
||||||
n, m := decodePair(x)
|
n, m := decodePair(x)
|
||||||
|
|
|
||||||
|
|
@ -1022,16 +1022,7 @@ func TestEIP8024_Execution(t *testing.T) {
|
||||||
}{
|
}{
|
||||||
{
|
{
|
||||||
name: "DUPN",
|
name: "DUPN",
|
||||||
codeHex: "60016000808080808080808080808080808080e600",
|
codeHex: "60016000808080808080808080808080808080e680",
|
||||||
wantVals: []uint64{
|
|
||||||
1,
|
|
||||||
0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
|
|
||||||
1,
|
|
||||||
},
|
|
||||||
},
|
|
||||||
{
|
|
||||||
name: "DUPN_MISSING_IMMEDIATE",
|
|
||||||
codeHex: "60016000808080808080808080808080808080e6",
|
|
||||||
wantVals: []uint64{
|
wantVals: []uint64{
|
||||||
1,
|
1,
|
||||||
0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
|
0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
|
||||||
|
|
@ -1040,7 +1031,7 @@ func TestEIP8024_Execution(t *testing.T) {
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
name: "SWAPN",
|
name: "SWAPN",
|
||||||
codeHex: "600160008080808080808080808080808080806002e700",
|
codeHex: "600160008080808080808080808080808080806002e780",
|
||||||
wantVals: []uint64{
|
wantVals: []uint64{
|
||||||
1,
|
1,
|
||||||
0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
|
0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
|
||||||
|
|
@ -1048,22 +1039,23 @@ func TestEIP8024_Execution(t *testing.T) {
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
name: "SWAPN_MISSING_IMMEDIATE",
|
name: "EXCHANGE_MISSING_IMMEDIATE",
|
||||||
codeHex: "600160008080808080808080808080808080806002e7",
|
codeHex: "600260008080808080600160008080808080808080e8",
|
||||||
wantVals: []uint64{
|
wantVals: []uint64{
|
||||||
1,
|
0, 0, 0, 0, 0, 0, 0, 0, 0,
|
||||||
0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
|
2, // 10th from top
|
||||||
2,
|
0, 0, 0, 0, 0, 0,
|
||||||
|
1, // bottom
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
name: "EXCHANGE",
|
name: "EXCHANGE",
|
||||||
codeHex: "600060016002e801",
|
codeHex: "600060016002e88e",
|
||||||
wantVals: []uint64{2, 0, 1},
|
wantVals: []uint64{2, 0, 1},
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
name: "EXCHANGE_MISSING_IMMEDIATE",
|
name: "EXCHANGE",
|
||||||
codeHex: "600060006000600060006000600060006000600060006000600060006000600060006000600060006000600060006000600060006000600060016002e8",
|
codeHex: "600080808080808080808080808080808080808080808080808080808060016002e88f",
|
||||||
wantVals: []uint64{
|
wantVals: []uint64{
|
||||||
2,
|
2,
|
||||||
0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
|
0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
|
||||||
|
|
@ -1077,68 +1069,31 @@ func TestEIP8024_Execution(t *testing.T) {
|
||||||
wantOpcode: SWAPN,
|
wantOpcode: SWAPN,
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
name: "JUMP over INVALID_DUPN",
|
name: "JUMP_OVER_INVALID_DUPN",
|
||||||
codeHex: "600456e65b",
|
codeHex: "600456e65b",
|
||||||
wantErr: nil,
|
wantErr: nil,
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
name: "UNDERFLOW_DUPN_1",
|
name: "EXCHANGE",
|
||||||
codeHex: "6000808080808080808080808080808080e600",
|
codeHex: "60008080e88e15",
|
||||||
|
wantVals: []uint64{1, 0, 0},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "INVALID_EXCHANGE",
|
||||||
|
codeHex: "e852",
|
||||||
|
wantErr: &ErrInvalidOpCode{},
|
||||||
|
wantOpcode: EXCHANGE,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "UNDERFLOW_DUPN",
|
||||||
|
codeHex: "6000808080808080808080808080808080e680",
|
||||||
wantErr: &ErrStackUnderflow{},
|
wantErr: &ErrStackUnderflow{},
|
||||||
wantOpcode: DUPN,
|
wantOpcode: DUPN,
|
||||||
},
|
},
|
||||||
// Additional test cases
|
// Additional test cases
|
||||||
{
|
|
||||||
name: "INVALID_DUPN_LOW",
|
|
||||||
codeHex: "e65b",
|
|
||||||
wantErr: &ErrInvalidOpCode{},
|
|
||||||
wantOpcode: DUPN,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
name: "INVALID_EXCHANGE_LOW",
|
|
||||||
codeHex: "e850",
|
|
||||||
wantErr: &ErrInvalidOpCode{},
|
|
||||||
wantOpcode: EXCHANGE,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
name: "INVALID_DUPN_HIGH",
|
|
||||||
codeHex: "e67f",
|
|
||||||
wantErr: &ErrInvalidOpCode{},
|
|
||||||
wantOpcode: DUPN,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
name: "INVALID_SWAPN_HIGH",
|
|
||||||
codeHex: "e77f",
|
|
||||||
wantErr: &ErrInvalidOpCode{},
|
|
||||||
wantOpcode: SWAPN,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
name: "INVALID_EXCHANGE_HIGH",
|
|
||||||
codeHex: "e87f",
|
|
||||||
wantErr: &ErrInvalidOpCode{},
|
|
||||||
wantOpcode: EXCHANGE,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
name: "UNDERFLOW_DUPN_2",
|
|
||||||
codeHex: "5f5f5f5f5f5f5f5f5f5f5f5f5f5f5f5fe600", // (n=17, need 17 items, have 16)
|
|
||||||
wantErr: &ErrStackUnderflow{},
|
|
||||||
wantOpcode: DUPN,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
name: "UNDERFLOW_SWAPN",
|
|
||||||
codeHex: "5f5f5f5f5f5f5f5f5f5f5f5f5f5f5f5f5fe700", // (n=17, need 18 items, have 17)
|
|
||||||
wantErr: &ErrStackUnderflow{},
|
|
||||||
wantOpcode: SWAPN,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
name: "UNDERFLOW_EXCHANGE",
|
|
||||||
codeHex: "60016002e801", // (n,m)=(1,2), need 3 items, have 2
|
|
||||||
wantErr: &ErrStackUnderflow{},
|
|
||||||
wantOpcode: EXCHANGE,
|
|
||||||
},
|
|
||||||
{
|
{
|
||||||
name: "PC_INCREMENT",
|
name: "PC_INCREMENT",
|
||||||
codeHex: "600060006000e80115",
|
codeHex: "600060006000e88e15",
|
||||||
wantVals: []uint64{1, 0, 0},
|
wantVals: []uint64{1, 0, 0},
|
||||||
},
|
},
|
||||||
}
|
}
|
||||||
|
|
|
||||||
|
|
@ -27,13 +27,11 @@ import (
|
||||||
|
|
||||||
// Config are the configuration options for the Interpreter
|
// Config are the configuration options for the Interpreter
|
||||||
type Config struct {
|
type Config struct {
|
||||||
Tracer *tracing.Hooks
|
Tracer *tracing.Hooks
|
||||||
|
|
||||||
NoBaseFee bool // Forces the EIP-1559 baseFee to 0 (needed for 0 price calls)
|
NoBaseFee bool // Forces the EIP-1559 baseFee to 0 (needed for 0 price calls)
|
||||||
EnablePreimageRecording bool // Enables recording of SHA3/keccak preimages
|
EnablePreimageRecording bool // Enables recording of SHA3/keccak preimages
|
||||||
ExtraEips []int // Additional EIPS that are to be enabled
|
ExtraEips []int // Additional EIPS that are to be enabled
|
||||||
|
|
||||||
StatelessSelfValidation bool // Generate execution witnesses and self-check against them (testing purpose)
|
|
||||||
EnableWitnessStats bool // Whether trie access statistics collection is enabled
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// ScopeContext contains the things that are per-call, such as stack and memory,
|
// ScopeContext contains the things that are per-call, such as stack and memory,
|
||||||
|
|
|
||||||
|
|
@ -63,6 +63,7 @@ var (
|
||||||
verkleInstructionSet = newVerkleInstructionSet()
|
verkleInstructionSet = newVerkleInstructionSet()
|
||||||
pragueInstructionSet = newPragueInstructionSet()
|
pragueInstructionSet = newPragueInstructionSet()
|
||||||
osakaInstructionSet = newOsakaInstructionSet()
|
osakaInstructionSet = newOsakaInstructionSet()
|
||||||
|
amsterdamInstructionSet = newAmsterdamInstructionSet()
|
||||||
)
|
)
|
||||||
|
|
||||||
// JumpTable contains the EVM opcodes supported at a given fork.
|
// JumpTable contains the EVM opcodes supported at a given fork.
|
||||||
|
|
@ -92,6 +93,13 @@ func newVerkleInstructionSet() JumpTable {
|
||||||
return validate(instructionSet)
|
return validate(instructionSet)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func newAmsterdamInstructionSet() JumpTable {
|
||||||
|
instructionSet := newOsakaInstructionSet()
|
||||||
|
enable7843(&instructionSet) // EIP-7843 (SLOTNUM opcode)
|
||||||
|
enable8024(&instructionSet) // EIP-8024 (Backward compatible SWAPN, DUPN, EXCHANGE)
|
||||||
|
return validate(instructionSet)
|
||||||
|
}
|
||||||
|
|
||||||
func newOsakaInstructionSet() JumpTable {
|
func newOsakaInstructionSet() JumpTable {
|
||||||
instructionSet := newPragueInstructionSet()
|
instructionSet := newPragueInstructionSet()
|
||||||
enable7939(&instructionSet) // EIP-7939 (CLZ opcode)
|
enable7939(&instructionSet) // EIP-7939 (CLZ opcode)
|
||||||
|
|
|
||||||
|
|
@ -28,6 +28,8 @@ func LookupInstructionSet(rules params.Rules) (JumpTable, error) {
|
||||||
switch {
|
switch {
|
||||||
case rules.IsVerkle:
|
case rules.IsVerkle:
|
||||||
return newCancunInstructionSet(), errors.New("verkle-fork not defined yet")
|
return newCancunInstructionSet(), errors.New("verkle-fork not defined yet")
|
||||||
|
case rules.IsAmsterdam:
|
||||||
|
return newAmsterdamInstructionSet(), nil
|
||||||
case rules.IsOsaka:
|
case rules.IsOsaka:
|
||||||
return newOsakaInstructionSet(), nil
|
return newOsakaInstructionSet(), nil
|
||||||
case rules.IsPrague:
|
case rules.IsPrague:
|
||||||
|
|
|
||||||
|
|
@ -105,6 +105,7 @@ const (
|
||||||
BASEFEE OpCode = 0x48
|
BASEFEE OpCode = 0x48
|
||||||
BLOBHASH OpCode = 0x49
|
BLOBHASH OpCode = 0x49
|
||||||
BLOBBASEFEE OpCode = 0x4a
|
BLOBBASEFEE OpCode = 0x4a
|
||||||
|
SLOTNUM OpCode = 0x4b
|
||||||
)
|
)
|
||||||
|
|
||||||
// 0x50 range - 'storage' and execution.
|
// 0x50 range - 'storage' and execution.
|
||||||
|
|
@ -320,6 +321,7 @@ var opCodeToString = [256]string{
|
||||||
BASEFEE: "BASEFEE",
|
BASEFEE: "BASEFEE",
|
||||||
BLOBHASH: "BLOBHASH",
|
BLOBHASH: "BLOBHASH",
|
||||||
BLOBBASEFEE: "BLOBBASEFEE",
|
BLOBBASEFEE: "BLOBBASEFEE",
|
||||||
|
SLOTNUM: "SLOTNUM",
|
||||||
|
|
||||||
// 0x50 range - 'storage' and execution.
|
// 0x50 range - 'storage' and execution.
|
||||||
POP: "POP",
|
POP: "POP",
|
||||||
|
|
@ -502,6 +504,7 @@ var stringToOp = map[string]OpCode{
|
||||||
"BASEFEE": BASEFEE,
|
"BASEFEE": BASEFEE,
|
||||||
"BLOBHASH": BLOBHASH,
|
"BLOBHASH": BLOBHASH,
|
||||||
"BLOBBASEFEE": BLOBBASEFEE,
|
"BLOBBASEFEE": BLOBBASEFEE,
|
||||||
|
"SLOTNUM": SLOTNUM,
|
||||||
"DELEGATECALL": DELEGATECALL,
|
"DELEGATECALL": DELEGATECALL,
|
||||||
"STATICCALL": STATICCALL,
|
"STATICCALL": STATICCALL,
|
||||||
"CODESIZE": CODESIZE,
|
"CODESIZE": CODESIZE,
|
||||||
|
|
|
||||||
|
|
@ -24,6 +24,7 @@ import (
|
||||||
|
|
||||||
"github.com/ethereum/go-ethereum/common"
|
"github.com/ethereum/go-ethereum/common"
|
||||||
"github.com/ethereum/go-ethereum/common/hexutil"
|
"github.com/ethereum/go-ethereum/common/hexutil"
|
||||||
|
"github.com/ethereum/go-ethereum/core"
|
||||||
"github.com/ethereum/go-ethereum/core/rawdb"
|
"github.com/ethereum/go-ethereum/core/rawdb"
|
||||||
"github.com/ethereum/go-ethereum/core/state"
|
"github.com/ethereum/go-ethereum/core/state"
|
||||||
"github.com/ethereum/go-ethereum/core/stateless"
|
"github.com/ethereum/go-ethereum/core/stateless"
|
||||||
|
|
@ -493,34 +494,22 @@ func (api *DebugAPI) StateSize(blockHashOrNumber *rpc.BlockNumberOrHash) (interf
|
||||||
}, nil
|
}, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func (api *DebugAPI) ExecutionWitness(bn rpc.BlockNumber) (*stateless.ExtWitness, error) {
|
func (api *DebugAPI) ExecutionWitness(bn rpc.BlockNumberOrHash) (*stateless.ExtWitness, error) {
|
||||||
bc := api.eth.blockchain
|
bc := api.eth.blockchain
|
||||||
block, err := api.eth.APIBackend.BlockByNumber(context.Background(), bn)
|
block, err := api.eth.APIBackend.BlockByNumberOrHash(context.Background(), bn)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return &stateless.ExtWitness{}, fmt.Errorf("block number %v not found", bn)
|
return &stateless.ExtWitness{}, fmt.Errorf("block %v not found", bn)
|
||||||
}
|
}
|
||||||
parent := bc.GetHeader(block.ParentHash(), block.NumberU64()-1)
|
parent := bc.GetHeader(block.ParentHash(), block.NumberU64()-1)
|
||||||
if parent == nil {
|
if parent == nil {
|
||||||
return &stateless.ExtWitness{}, fmt.Errorf("block number %v found, but parent missing", bn)
|
return &stateless.ExtWitness{}, fmt.Errorf("block %v found, but parent missing", bn)
|
||||||
}
|
}
|
||||||
result, err := bc.ProcessBlock(context.Background(), parent.Root, block, false, true)
|
config := core.ExecuteConfig{
|
||||||
if err != nil {
|
WriteState: false,
|
||||||
return nil, err
|
EnableTracer: false,
|
||||||
}
|
MakeWitness: true,
|
||||||
return result.Witness().ToExtWitness(), nil
|
}
|
||||||
}
|
result, err := bc.ProcessBlock(context.Background(), parent.Root, block, config)
|
||||||
|
|
||||||
func (api *DebugAPI) ExecutionWitnessByHash(hash common.Hash) (*stateless.ExtWitness, error) {
|
|
||||||
bc := api.eth.blockchain
|
|
||||||
block := bc.GetBlockByHash(hash)
|
|
||||||
if block == nil {
|
|
||||||
return &stateless.ExtWitness{}, fmt.Errorf("block hash %x not found", hash)
|
|
||||||
}
|
|
||||||
parent := bc.GetHeader(block.ParentHash(), block.NumberU64()-1)
|
|
||||||
if parent == nil {
|
|
||||||
return &stateless.ExtWitness{}, fmt.Errorf("block number %x found, but parent missing", hash)
|
|
||||||
}
|
|
||||||
result, err := bc.ProcessBlock(context.Background(), parent.Root, block, false, true)
|
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
|
|
|
||||||
|
|
@ -237,8 +237,6 @@ func New(stack *node.Node, config *ethconfig.Config) (*Ethereum, error) {
|
||||||
TxLookupLimit: int64(min(config.TransactionHistory, math.MaxInt64)),
|
TxLookupLimit: int64(min(config.TransactionHistory, math.MaxInt64)),
|
||||||
VmConfig: vm.Config{
|
VmConfig: vm.Config{
|
||||||
EnablePreimageRecording: config.EnablePreimageRecording,
|
EnablePreimageRecording: config.EnablePreimageRecording,
|
||||||
EnableWitnessStats: config.EnableWitnessStats,
|
|
||||||
StatelessSelfValidation: config.StatelessSelfValidation,
|
|
||||||
},
|
},
|
||||||
// Enables file journaling for the trie database. The journal files will be stored
|
// Enables file journaling for the trie database. The journal files will be stored
|
||||||
// within the data directory. The corresponding paths will be either:
|
// within the data directory. The corresponding paths will be either:
|
||||||
|
|
@ -247,6 +245,9 @@ func New(stack *node.Node, config *ethconfig.Config) (*Ethereum, error) {
|
||||||
TrieJournalDirectory: stack.ResolvePath("triedb"),
|
TrieJournalDirectory: stack.ResolvePath("triedb"),
|
||||||
StateSizeTracking: config.EnableStateSizeTracking,
|
StateSizeTracking: config.EnableStateSizeTracking,
|
||||||
SlowBlockThreshold: config.SlowBlockThreshold,
|
SlowBlockThreshold: config.SlowBlockThreshold,
|
||||||
|
|
||||||
|
StatelessSelfValidation: config.StatelessSelfValidation,
|
||||||
|
EnableWitnessStats: config.EnableWitnessStats,
|
||||||
}
|
}
|
||||||
)
|
)
|
||||||
if config.VMTrace != "" {
|
if config.VMTrace != "" {
|
||||||
|
|
|
||||||
|
|
@ -47,9 +47,10 @@ import (
|
||||||
"github.com/ethereum/go-ethereum/rpc"
|
"github.com/ethereum/go-ethereum/rpc"
|
||||||
)
|
)
|
||||||
|
|
||||||
// Register adds the engine API to the full node.
|
// Register adds the engine API and related APIs to the full node.
|
||||||
func Register(stack *node.Node, backend *eth.Ethereum) error {
|
func Register(stack *node.Node, backend *eth.Ethereum) error {
|
||||||
stack.RegisterAPIs([]rpc.API{
|
stack.RegisterAPIs([]rpc.API{
|
||||||
|
newTestingAPI(backend),
|
||||||
{
|
{
|
||||||
Namespace: "engine",
|
Namespace: "engine",
|
||||||
Service: NewConsensusAPI(backend),
|
Service: NewConsensusAPI(backend),
|
||||||
|
|
@ -211,6 +212,28 @@ func (api *ConsensusAPI) ForkchoiceUpdatedV3(update engine.ForkchoiceStateV1, pa
|
||||||
return api.forkchoiceUpdated(update, params, engine.PayloadV3, false)
|
return api.forkchoiceUpdated(update, params, engine.PayloadV3, false)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// ForkchoiceUpdatedV4 is equivalent to V3 with the addition of slot number
|
||||||
|
// in the payload attributes. It supports only PayloadAttributesV4.
|
||||||
|
func (api *ConsensusAPI) ForkchoiceUpdatedV4(update engine.ForkchoiceStateV1, params *engine.PayloadAttributes) (engine.ForkChoiceResponse, error) {
|
||||||
|
if params != nil {
|
||||||
|
switch {
|
||||||
|
case params.Withdrawals == nil:
|
||||||
|
return engine.STATUS_INVALID, attributesErr("missing withdrawals")
|
||||||
|
case params.BeaconRoot == nil:
|
||||||
|
return engine.STATUS_INVALID, attributesErr("missing beacon root")
|
||||||
|
case params.SlotNumber == nil:
|
||||||
|
return engine.STATUS_INVALID, attributesErr("missing slot number")
|
||||||
|
case !api.checkFork(params.Timestamp, forks.Amsterdam):
|
||||||
|
return engine.STATUS_INVALID, unsupportedForkErr("fcuV4 must only be called for amsterdam payloads")
|
||||||
|
}
|
||||||
|
}
|
||||||
|
// TODO(matt): the spec requires that fcu is applied when called on a valid
|
||||||
|
// hash, even if params are wrong. To do this we need to split up
|
||||||
|
// forkchoiceUpdate into a function that only updates the head and then a
|
||||||
|
// function that kicks off block construction.
|
||||||
|
return api.forkchoiceUpdated(update, params, engine.PayloadV4, false)
|
||||||
|
}
|
||||||
|
|
||||||
func (api *ConsensusAPI) forkchoiceUpdated(update engine.ForkchoiceStateV1, payloadAttributes *engine.PayloadAttributes, payloadVersion engine.PayloadVersion, payloadWitness bool) (engine.ForkChoiceResponse, error) {
|
func (api *ConsensusAPI) forkchoiceUpdated(update engine.ForkchoiceStateV1, payloadAttributes *engine.PayloadAttributes, payloadVersion engine.PayloadVersion, payloadWitness bool) (engine.ForkChoiceResponse, error) {
|
||||||
api.forkchoiceLock.Lock()
|
api.forkchoiceLock.Lock()
|
||||||
defer api.forkchoiceLock.Unlock()
|
defer api.forkchoiceLock.Unlock()
|
||||||
|
|
@ -344,6 +367,7 @@ func (api *ConsensusAPI) forkchoiceUpdated(update engine.ForkchoiceStateV1, payl
|
||||||
Random: payloadAttributes.Random,
|
Random: payloadAttributes.Random,
|
||||||
Withdrawals: payloadAttributes.Withdrawals,
|
Withdrawals: payloadAttributes.Withdrawals,
|
||||||
BeaconRoot: payloadAttributes.BeaconRoot,
|
BeaconRoot: payloadAttributes.BeaconRoot,
|
||||||
|
SlotNum: payloadAttributes.SlotNumber,
|
||||||
Version: payloadVersion,
|
Version: payloadVersion,
|
||||||
}
|
}
|
||||||
id := args.Id()
|
id := args.Id()
|
||||||
|
|
@ -457,6 +481,18 @@ func (api *ConsensusAPI) GetPayloadV5(payloadID engine.PayloadID) (*engine.Execu
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// GetPayloadV6 returns a cached payload by id. This endpoint should only
|
||||||
|
// be used after the Amsterdam fork.
|
||||||
|
func (api *ConsensusAPI) GetPayloadV6(payloadID engine.PayloadID) (*engine.ExecutionPayloadEnvelope, error) {
|
||||||
|
return api.getPayload(
|
||||||
|
payloadID,
|
||||||
|
false,
|
||||||
|
[]engine.PayloadVersion{engine.PayloadV4},
|
||||||
|
[]forks.Fork{
|
||||||
|
forks.Amsterdam,
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
// getPayload will retrieve the specified payload and verify it conforms to the
|
// getPayload will retrieve the specified payload and verify it conforms to the
|
||||||
// endpoint's allowed payload versions and forks.
|
// endpoint's allowed payload versions and forks.
|
||||||
//
|
//
|
||||||
|
|
@ -699,6 +735,33 @@ func (api *ConsensusAPI) NewPayloadV4(ctx context.Context, params engine.Executa
|
||||||
return api.newPayload(ctx, params, versionedHashes, beaconRoot, requests, false)
|
return api.newPayload(ctx, params, versionedHashes, beaconRoot, requests, false)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// NewPayloadV5 creates an Eth1 block, inserts it in the chain, and returns the status of the chain.
|
||||||
|
func (api *ConsensusAPI) NewPayloadV5(ctx context.Context, params engine.ExecutableData, versionedHashes []common.Hash, beaconRoot *common.Hash, executionRequests []hexutil.Bytes) (engine.PayloadStatusV1, error) {
|
||||||
|
switch {
|
||||||
|
case params.Withdrawals == nil:
|
||||||
|
return invalidStatus, paramsErr("nil withdrawals post-shanghai")
|
||||||
|
case params.ExcessBlobGas == nil:
|
||||||
|
return invalidStatus, paramsErr("nil excessBlobGas post-cancun")
|
||||||
|
case params.BlobGasUsed == nil:
|
||||||
|
return invalidStatus, paramsErr("nil blobGasUsed post-cancun")
|
||||||
|
case versionedHashes == nil:
|
||||||
|
return invalidStatus, paramsErr("nil versionedHashes post-cancun")
|
||||||
|
case beaconRoot == nil:
|
||||||
|
return invalidStatus, paramsErr("nil beaconRoot post-cancun")
|
||||||
|
case executionRequests == nil:
|
||||||
|
return invalidStatus, paramsErr("nil executionRequests post-prague")
|
||||||
|
case params.SlotNumber == nil:
|
||||||
|
return invalidStatus, paramsErr("nil slotnumber post-amsterdam")
|
||||||
|
case !api.checkFork(params.Timestamp, forks.Amsterdam):
|
||||||
|
return invalidStatus, unsupportedForkErr("newPayloadV5 must only be called for amsterdam payloads")
|
||||||
|
}
|
||||||
|
requests := convertRequests(executionRequests)
|
||||||
|
if err := validateRequests(requests); err != nil {
|
||||||
|
return engine.PayloadStatusV1{Status: engine.INVALID}, engine.InvalidParams.With(err)
|
||||||
|
}
|
||||||
|
return api.newPayload(ctx, params, versionedHashes, beaconRoot, requests, false)
|
||||||
|
}
|
||||||
|
|
||||||
func (api *ConsensusAPI) newPayload(ctx context.Context, params engine.ExecutableData, versionedHashes []common.Hash, beaconRoot *common.Hash, requests [][]byte, witness bool) (result engine.PayloadStatusV1, err error) {
|
func (api *ConsensusAPI) newPayload(ctx context.Context, params engine.ExecutableData, versionedHashes []common.Hash, beaconRoot *common.Hash, requests [][]byte, witness bool) (result engine.PayloadStatusV1, err error) {
|
||||||
// The locking here is, strictly, not required. Without these locks, this can happen:
|
// The locking here is, strictly, not required. Without these locks, this can happen:
|
||||||
//
|
//
|
||||||
|
|
@ -734,6 +797,10 @@ func (api *ConsensusAPI) newPayload(ctx context.Context, params engine.Executabl
|
||||||
if params.ExcessBlobGas != nil {
|
if params.ExcessBlobGas != nil {
|
||||||
ebg = strconv.Itoa(int(*params.ExcessBlobGas))
|
ebg = strconv.Itoa(int(*params.ExcessBlobGas))
|
||||||
}
|
}
|
||||||
|
slotNum := "nil"
|
||||||
|
if params.SlotNumber != nil {
|
||||||
|
slotNum = strconv.Itoa(int(*params.SlotNumber))
|
||||||
|
}
|
||||||
log.Warn("Invalid NewPayload params",
|
log.Warn("Invalid NewPayload params",
|
||||||
"params.Number", params.Number,
|
"params.Number", params.Number,
|
||||||
"params.ParentHash", params.ParentHash,
|
"params.ParentHash", params.ParentHash,
|
||||||
|
|
@ -749,6 +816,7 @@ func (api *ConsensusAPI) newPayload(ctx context.Context, params engine.Executabl
|
||||||
"params.BaseFeePerGas", params.BaseFeePerGas,
|
"params.BaseFeePerGas", params.BaseFeePerGas,
|
||||||
"params.BlobGasUsed", bgu,
|
"params.BlobGasUsed", bgu,
|
||||||
"params.ExcessBlobGas", ebg,
|
"params.ExcessBlobGas", ebg,
|
||||||
|
"params.SlotNumber", slotNum,
|
||||||
"len(params.Transactions)", len(params.Transactions),
|
"len(params.Transactions)", len(params.Transactions),
|
||||||
"len(params.Withdrawals)", len(params.Withdrawals),
|
"len(params.Withdrawals)", len(params.Withdrawals),
|
||||||
"beaconRoot", beaconRoot,
|
"beaconRoot", beaconRoot,
|
||||||
|
|
|
||||||
79
eth/catalyst/api_testing.go
Normal file
79
eth/catalyst/api_testing.go
Normal file
|
|
@ -0,0 +1,79 @@
|
||||||
|
// Copyright 2026 The go-ethereum Authors
|
||||||
|
// This file is part of the go-ethereum library.
|
||||||
|
//
|
||||||
|
// The go-ethereum library is free software: you can redistribute it and/or modify
|
||||||
|
// it under the terms of the GNU Lesser General Public License as published by
|
||||||
|
// the Free Software Foundation, either version 3 of the License, or
|
||||||
|
// (at your option) any later version.
|
||||||
|
//
|
||||||
|
// The go-ethereum library is distributed in the hope that it will be useful,
|
||||||
|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
||||||
|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
||||||
|
// GNU Lesser General Public License for more details.
|
||||||
|
//
|
||||||
|
// You should have received a copy of the GNU Lesser General Public License
|
||||||
|
// along with the go-ethereum library. If not, see <http://www.gnu.org/licenses/>.
|
||||||
|
|
||||||
|
package catalyst
|
||||||
|
|
||||||
|
import (
|
||||||
|
"errors"
|
||||||
|
|
||||||
|
"github.com/ethereum/go-ethereum/beacon/engine"
|
||||||
|
"github.com/ethereum/go-ethereum/common"
|
||||||
|
"github.com/ethereum/go-ethereum/common/hexutil"
|
||||||
|
"github.com/ethereum/go-ethereum/core/types"
|
||||||
|
"github.com/ethereum/go-ethereum/eth"
|
||||||
|
"github.com/ethereum/go-ethereum/miner"
|
||||||
|
"github.com/ethereum/go-ethereum/rpc"
|
||||||
|
)
|
||||||
|
|
||||||
|
// testingAPI implements the testing_ namespace.
|
||||||
|
// It's an engine-API adjacent namespace for testing purposes.
|
||||||
|
type testingAPI struct {
|
||||||
|
eth *eth.Ethereum
|
||||||
|
}
|
||||||
|
|
||||||
|
func newTestingAPI(backend *eth.Ethereum) rpc.API {
|
||||||
|
return rpc.API{
|
||||||
|
Namespace: "testing",
|
||||||
|
Service: &testingAPI{backend},
|
||||||
|
Version: "1.0",
|
||||||
|
Authenticated: false,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func (api *testingAPI) BuildBlockV1(parentHash common.Hash, payloadAttributes engine.PayloadAttributes, transactions *[]hexutil.Bytes, extraData *hexutil.Bytes) (*engine.ExecutionPayloadEnvelope, error) {
|
||||||
|
if api.eth.BlockChain().CurrentBlock().Hash() != parentHash {
|
||||||
|
return nil, errors.New("parentHash is not current head")
|
||||||
|
}
|
||||||
|
// If transactions is empty but not nil, build an empty block
|
||||||
|
// If the transactions is nil, build a block with the current transactions from the txpool
|
||||||
|
// If the transactions is not nil and not empty, build a block with the transactions
|
||||||
|
buildEmpty := transactions != nil && len(*transactions) == 0
|
||||||
|
var txs []*types.Transaction
|
||||||
|
if transactions != nil {
|
||||||
|
dec := make([][]byte, 0, len(*transactions))
|
||||||
|
for _, tx := range *transactions {
|
||||||
|
dec = append(dec, tx)
|
||||||
|
}
|
||||||
|
var err error
|
||||||
|
txs, err = engine.DecodeTransactions(dec)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
}
|
||||||
|
extra := make([]byte, 0)
|
||||||
|
if extraData != nil {
|
||||||
|
extra = *extraData
|
||||||
|
}
|
||||||
|
args := &miner.BuildPayloadArgs{
|
||||||
|
Parent: parentHash,
|
||||||
|
Timestamp: payloadAttributes.Timestamp,
|
||||||
|
FeeRecipient: payloadAttributes.SuggestedFeeRecipient,
|
||||||
|
Random: payloadAttributes.Random,
|
||||||
|
Withdrawals: payloadAttributes.Withdrawals,
|
||||||
|
BeaconRoot: payloadAttributes.BeaconRoot,
|
||||||
|
}
|
||||||
|
return api.eth.Miner().BuildTestingPayload(args, txs, buildEmpty, extra)
|
||||||
|
}
|
||||||
121
eth/catalyst/api_testing_test.go
Normal file
121
eth/catalyst/api_testing_test.go
Normal file
|
|
@ -0,0 +1,121 @@
|
||||||
|
// Copyright 2026 The go-ethereum Authors
|
||||||
|
// This file is part of the go-ethereum library.
|
||||||
|
//
|
||||||
|
// The go-ethereum library is free software: you can redistribute it and/or modify
|
||||||
|
// it under the terms of the GNU Lesser General Public License as published by
|
||||||
|
// the Free Software Foundation, either version 3 of the License, or
|
||||||
|
// (at your option) any later version.
|
||||||
|
//
|
||||||
|
// The go-ethereum library is distributed in the hope that it will be useful,
|
||||||
|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
||||||
|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
||||||
|
// GNU Lesser General Public License for more details.
|
||||||
|
//
|
||||||
|
// You should have received a copy of the GNU Lesser General Public License
|
||||||
|
// along with the go-ethereum library. If not, see <http://www.gnu.org/licenses/>.
|
||||||
|
|
||||||
|
package catalyst
|
||||||
|
|
||||||
|
import (
|
||||||
|
"context"
|
||||||
|
"math/big"
|
||||||
|
"testing"
|
||||||
|
|
||||||
|
"github.com/ethereum/go-ethereum/beacon/engine"
|
||||||
|
"github.com/ethereum/go-ethereum/common"
|
||||||
|
"github.com/ethereum/go-ethereum/common/hexutil"
|
||||||
|
"github.com/ethereum/go-ethereum/core/types"
|
||||||
|
"github.com/ethereum/go-ethereum/crypto"
|
||||||
|
"github.com/ethereum/go-ethereum/params"
|
||||||
|
)
|
||||||
|
|
||||||
|
func TestBuildBlockV1(t *testing.T) {
|
||||||
|
genesis, blocks := generateMergeChain(5, true)
|
||||||
|
n, ethservice := startEthService(t, genesis, blocks)
|
||||||
|
defer n.Close()
|
||||||
|
|
||||||
|
parent := ethservice.BlockChain().CurrentBlock()
|
||||||
|
attrs := engine.PayloadAttributes{
|
||||||
|
Timestamp: parent.Time + 1,
|
||||||
|
Random: crypto.Keccak256Hash([]byte("test")),
|
||||||
|
SuggestedFeeRecipient: parent.Coinbase,
|
||||||
|
Withdrawals: nil,
|
||||||
|
BeaconRoot: nil,
|
||||||
|
}
|
||||||
|
|
||||||
|
currentNonce, _ := ethservice.APIBackend.GetPoolNonce(context.Background(), testAddr)
|
||||||
|
tx, _ := types.SignTx(types.NewTransaction(currentNonce, testAddr, big.NewInt(1), params.TxGas, big.NewInt(params.InitialBaseFee*2), nil), types.LatestSigner(ethservice.BlockChain().Config()), testKey)
|
||||||
|
|
||||||
|
api := &testingAPI{eth: ethservice}
|
||||||
|
|
||||||
|
t.Run("buildOnCurrentHead", func(t *testing.T) {
|
||||||
|
envelope, err := api.BuildBlockV1(parent.Hash(), attrs, nil, nil)
|
||||||
|
if err != nil {
|
||||||
|
t.Fatalf("BuildBlockV1 failed: %v", err)
|
||||||
|
}
|
||||||
|
if envelope == nil || envelope.ExecutionPayload == nil {
|
||||||
|
t.Fatal("expected non-nil envelope and payload")
|
||||||
|
}
|
||||||
|
payload := envelope.ExecutionPayload
|
||||||
|
if payload.ParentHash != parent.Hash() {
|
||||||
|
t.Errorf("parent hash mismatch: got %x want %x", payload.ParentHash, parent.Hash())
|
||||||
|
}
|
||||||
|
if payload.Number != parent.Number.Uint64()+1 {
|
||||||
|
t.Errorf("block number mismatch: got %d want %d", payload.Number, parent.Number.Uint64()+1)
|
||||||
|
}
|
||||||
|
if payload.Timestamp != attrs.Timestamp {
|
||||||
|
t.Errorf("timestamp mismatch: got %d want %d", payload.Timestamp, attrs.Timestamp)
|
||||||
|
}
|
||||||
|
if payload.FeeRecipient != attrs.SuggestedFeeRecipient {
|
||||||
|
t.Errorf("fee recipient mismatch: got %x want %x", payload.FeeRecipient, attrs.SuggestedFeeRecipient)
|
||||||
|
}
|
||||||
|
})
|
||||||
|
|
||||||
|
t.Run("wrongParentHash", func(t *testing.T) {
|
||||||
|
wrongParent := common.Hash{0x01}
|
||||||
|
_, err := api.BuildBlockV1(wrongParent, attrs, nil, nil)
|
||||||
|
if err == nil {
|
||||||
|
t.Fatal("expected error when parentHash is not current head")
|
||||||
|
}
|
||||||
|
if err.Error() != "parentHash is not current head" {
|
||||||
|
t.Errorf("unexpected error: %v", err)
|
||||||
|
}
|
||||||
|
})
|
||||||
|
|
||||||
|
t.Run("buildEmptyBlock", func(t *testing.T) {
|
||||||
|
emptyTxs := []hexutil.Bytes{}
|
||||||
|
envelope, err := api.BuildBlockV1(parent.Hash(), attrs, &emptyTxs, nil)
|
||||||
|
if err != nil {
|
||||||
|
t.Fatalf("BuildBlockV1 with empty txs failed: %v", err)
|
||||||
|
}
|
||||||
|
if envelope == nil || envelope.ExecutionPayload == nil {
|
||||||
|
t.Fatal("expected non-nil envelope and payload")
|
||||||
|
}
|
||||||
|
if len(envelope.ExecutionPayload.Transactions) != 0 {
|
||||||
|
t.Errorf("expected empty block, got %d transactions", len(envelope.ExecutionPayload.Transactions))
|
||||||
|
}
|
||||||
|
})
|
||||||
|
|
||||||
|
t.Run("buildBlockWithTransactions", func(t *testing.T) {
|
||||||
|
enc, _ := tx.MarshalBinary()
|
||||||
|
txs := []hexutil.Bytes{enc}
|
||||||
|
envelope, err := api.BuildBlockV1(parent.Hash(), attrs, &txs, nil)
|
||||||
|
if err != nil {
|
||||||
|
t.Fatalf("BuildBlockV1 with transaction failed: %v", err)
|
||||||
|
}
|
||||||
|
if len(envelope.ExecutionPayload.Transactions) != 1 {
|
||||||
|
t.Errorf("expected 1 transaction, got %d", len(envelope.ExecutionPayload.Transactions))
|
||||||
|
}
|
||||||
|
})
|
||||||
|
|
||||||
|
t.Run("buildBlockWithTransactionsFromTxPool", func(t *testing.T) {
|
||||||
|
ethservice.TxPool().Add([]*types.Transaction{tx}, true)
|
||||||
|
envelope, err := api.BuildBlockV1(parent.Hash(), attrs, nil, nil)
|
||||||
|
if err != nil {
|
||||||
|
t.Fatalf("BuildBlockV1 with transaction failed: %v", err)
|
||||||
|
}
|
||||||
|
if len(envelope.ExecutionPayload.Transactions) != 1 {
|
||||||
|
t.Errorf("expected 1 transaction, got %d", len(envelope.ExecutionPayload.Transactions))
|
||||||
|
}
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
@ -103,6 +103,8 @@ type SimulatedBeacon struct {
|
||||||
|
|
||||||
func payloadVersion(config *params.ChainConfig, time uint64) engine.PayloadVersion {
|
func payloadVersion(config *params.ChainConfig, time uint64) engine.PayloadVersion {
|
||||||
switch config.LatestFork(time) {
|
switch config.LatestFork(time) {
|
||||||
|
case forks.Amsterdam:
|
||||||
|
return engine.PayloadV4
|
||||||
case forks.BPO5, forks.BPO4, forks.BPO3, forks.BPO2, forks.BPO1, forks.Osaka, forks.Prague, forks.Cancun:
|
case forks.BPO5, forks.BPO4, forks.BPO3, forks.BPO2, forks.BPO1, forks.Osaka, forks.Prague, forks.Cancun:
|
||||||
return engine.PayloadV3
|
return engine.PayloadV3
|
||||||
case forks.Paris, forks.Shanghai:
|
case forks.Paris, forks.Shanghai:
|
||||||
|
|
@ -198,13 +200,19 @@ func (c *SimulatedBeacon) sealBlock(withdrawals []*types.Withdrawal, timestamp u
|
||||||
|
|
||||||
var random [32]byte
|
var random [32]byte
|
||||||
rand.Read(random[:])
|
rand.Read(random[:])
|
||||||
fcResponse, err := c.engineAPI.forkchoiceUpdated(c.curForkchoiceState, &engine.PayloadAttributes{
|
|
||||||
|
attribute := &engine.PayloadAttributes{
|
||||||
Timestamp: timestamp,
|
Timestamp: timestamp,
|
||||||
SuggestedFeeRecipient: feeRecipient,
|
SuggestedFeeRecipient: feeRecipient,
|
||||||
Withdrawals: withdrawals,
|
Withdrawals: withdrawals,
|
||||||
Random: random,
|
Random: random,
|
||||||
BeaconRoot: &common.Hash{},
|
BeaconRoot: &common.Hash{},
|
||||||
}, version, false)
|
}
|
||||||
|
if c.eth.BlockChain().Config().LatestFork(timestamp) == forks.Amsterdam {
|
||||||
|
slotNumber := uint64(0)
|
||||||
|
attribute.SlotNumber = &slotNumber
|
||||||
|
}
|
||||||
|
fcResponse, err := c.engineAPI.forkchoiceUpdated(c.curForkchoiceState, attribute, version, false)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
|
@ -376,5 +384,6 @@ func RegisterSimulatedBeaconAPIs(stack *node.Node, sim *SimulatedBeacon) {
|
||||||
Service: api,
|
Service: api,
|
||||||
Version: "1.0",
|
Version: "1.0",
|
||||||
},
|
},
|
||||||
|
newTestingAPI(sim.eth),
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
|
||||||
|
|
@ -17,7 +17,6 @@
|
||||||
package downloader
|
package downloader
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"fmt"
|
|
||||||
"math/big"
|
"math/big"
|
||||||
"sync"
|
"sync"
|
||||||
"sync/atomic"
|
"sync/atomic"
|
||||||
|
|
@ -261,23 +260,24 @@ func (dlp *downloadTesterPeer) RequestBodies(hashes []common.Hash, sink chan *et
|
||||||
// peer in the download tester. The returned function can be used to retrieve
|
// peer in the download tester. The returned function can be used to retrieve
|
||||||
// batches of block receipts from the particularly requested peer.
|
// batches of block receipts from the particularly requested peer.
|
||||||
func (dlp *downloadTesterPeer) RequestReceipts(hashes []common.Hash, sink chan *eth.Response) (*eth.Request, error) {
|
func (dlp *downloadTesterPeer) RequestReceipts(hashes []common.Hash, sink chan *eth.Response) (*eth.Request, error) {
|
||||||
blobs := eth.ServiceGetReceiptsQuery68(dlp.chain, hashes)
|
blobs := eth.ServiceGetReceiptsQuery(dlp.chain, hashes)
|
||||||
|
receipts := make([]types.Receipts, blobs.Len())
|
||||||
|
|
||||||
receipts := make([]types.Receipts, len(blobs))
|
// compute hashes
|
||||||
for i, blob := range blobs {
|
hashes = make([]common.Hash, blobs.Len())
|
||||||
rlp.DecodeBytes(blob, &receipts[i])
|
|
||||||
}
|
|
||||||
hasher := trie.NewStackTrie(nil)
|
hasher := trie.NewStackTrie(nil)
|
||||||
hashes = make([]common.Hash, len(receipts))
|
receiptLists, err := blobs.Items()
|
||||||
for i, receipt := range receipts {
|
if err != nil {
|
||||||
hashes[i] = types.DeriveSha(receipt, hasher)
|
panic(err)
|
||||||
}
|
}
|
||||||
req := ð.Request{
|
for i, rl := range receiptLists {
|
||||||
Peer: dlp.id,
|
hashes[i] = types.DeriveSha(rl.Derivable(), hasher)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// deliver the response right away
|
||||||
resp := eth.ReceiptsRLPResponse(types.EncodeBlockReceiptLists(receipts))
|
resp := eth.ReceiptsRLPResponse(types.EncodeBlockReceiptLists(receipts))
|
||||||
res := ð.Response{
|
res := ð.Response{
|
||||||
Req: req,
|
Req: ð.Request{Peer: dlp.id},
|
||||||
Res: &resp,
|
Res: &resp,
|
||||||
Meta: hashes,
|
Meta: hashes,
|
||||||
Time: 1,
|
Time: 1,
|
||||||
|
|
@ -286,7 +286,7 @@ func (dlp *downloadTesterPeer) RequestReceipts(hashes []common.Hash, sink chan *
|
||||||
go func() {
|
go func() {
|
||||||
sink <- res
|
sink <- res
|
||||||
}()
|
}()
|
||||||
return req, nil
|
return res.Req, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// ID retrieves the peer's unique identifier.
|
// ID retrieves the peer's unique identifier.
|
||||||
|
|
@ -398,8 +398,8 @@ func assertOwnChain(t *testing.T, tester *downloadTester, length int) {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestCanonicalSynchronisation68Full(t *testing.T) { testCanonSync(t, eth.ETH68, FullSync) }
|
func TestCanonicalSynchronisationFull(t *testing.T) { testCanonSync(t, eth.ETH69, FullSync) }
|
||||||
func TestCanonicalSynchronisation68Snap(t *testing.T) { testCanonSync(t, eth.ETH68, SnapSync) }
|
func TestCanonicalSynchronisationSnap(t *testing.T) { testCanonSync(t, eth.ETH69, SnapSync) }
|
||||||
|
|
||||||
func testCanonSync(t *testing.T, protocol uint, mode SyncMode) {
|
func testCanonSync(t *testing.T, protocol uint, mode SyncMode) {
|
||||||
success := make(chan struct{})
|
success := make(chan struct{})
|
||||||
|
|
@ -426,8 +426,8 @@ func testCanonSync(t *testing.T, protocol uint, mode SyncMode) {
|
||||||
|
|
||||||
// Tests that if a large batch of blocks are being downloaded, it is throttled
|
// Tests that if a large batch of blocks are being downloaded, it is throttled
|
||||||
// until the cached blocks are retrieved.
|
// until the cached blocks are retrieved.
|
||||||
func TestThrottling68Full(t *testing.T) { testThrottling(t, eth.ETH68, FullSync) }
|
func TestThrottlingFull(t *testing.T) { testThrottling(t, eth.ETH69, FullSync) }
|
||||||
func TestThrottling68Snap(t *testing.T) { testThrottling(t, eth.ETH68, SnapSync) }
|
func TestThrottlingSnap(t *testing.T) { testThrottling(t, eth.ETH69, SnapSync) }
|
||||||
|
|
||||||
func testThrottling(t *testing.T, protocol uint, mode SyncMode) {
|
func testThrottling(t *testing.T, protocol uint, mode SyncMode) {
|
||||||
tester := newTester(t, mode)
|
tester := newTester(t, mode)
|
||||||
|
|
@ -504,8 +504,8 @@ func testThrottling(t *testing.T, protocol uint, mode SyncMode) {
|
||||||
}
|
}
|
||||||
|
|
||||||
// Tests that a canceled download wipes all previously accumulated state.
|
// Tests that a canceled download wipes all previously accumulated state.
|
||||||
func TestCancel68Full(t *testing.T) { testCancel(t, eth.ETH68, FullSync) }
|
func TestCancelFull(t *testing.T) { testCancel(t, eth.ETH69, FullSync) }
|
||||||
func TestCancel68Snap(t *testing.T) { testCancel(t, eth.ETH68, SnapSync) }
|
func TestCancelSnap(t *testing.T) { testCancel(t, eth.ETH69, SnapSync) }
|
||||||
|
|
||||||
func testCancel(t *testing.T, protocol uint, mode SyncMode) {
|
func testCancel(t *testing.T, protocol uint, mode SyncMode) {
|
||||||
complete := make(chan struct{})
|
complete := make(chan struct{})
|
||||||
|
|
@ -534,49 +534,10 @@ func testCancel(t *testing.T, protocol uint, mode SyncMode) {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Tests that synchronisations behave well in multi-version protocol environments
|
|
||||||
// and not wreak havoc on other nodes in the network.
|
|
||||||
func TestMultiProtoSynchronisation68Full(t *testing.T) { testMultiProtoSync(t, eth.ETH68, FullSync) }
|
|
||||||
func TestMultiProtoSynchronisation68Snap(t *testing.T) { testMultiProtoSync(t, eth.ETH68, SnapSync) }
|
|
||||||
|
|
||||||
func testMultiProtoSync(t *testing.T, protocol uint, mode SyncMode) {
|
|
||||||
complete := make(chan struct{})
|
|
||||||
success := func() {
|
|
||||||
close(complete)
|
|
||||||
}
|
|
||||||
tester := newTesterWithNotification(t, mode, success)
|
|
||||||
defer tester.terminate()
|
|
||||||
|
|
||||||
// Create a small enough block chain to download
|
|
||||||
chain := testChainBase.shorten(blockCacheMaxItems - 15)
|
|
||||||
|
|
||||||
// Create peers of every type
|
|
||||||
tester.newPeer("peer 68", eth.ETH68, chain.blocks[1:])
|
|
||||||
|
|
||||||
if err := tester.downloader.BeaconSync(chain.blocks[len(chain.blocks)-1].Header(), nil); err != nil {
|
|
||||||
t.Fatalf("failed to start beacon sync: %v", err)
|
|
||||||
}
|
|
||||||
select {
|
|
||||||
case <-complete:
|
|
||||||
break
|
|
||||||
case <-time.NewTimer(time.Second * 3).C:
|
|
||||||
t.Fatalf("Failed to sync chain in three seconds")
|
|
||||||
}
|
|
||||||
assertOwnChain(t, tester, len(chain.blocks))
|
|
||||||
|
|
||||||
// Check that no peers have been dropped off
|
|
||||||
for _, version := range []int{68} {
|
|
||||||
peer := fmt.Sprintf("peer %d", version)
|
|
||||||
if _, ok := tester.peers[peer]; !ok {
|
|
||||||
t.Errorf("%s dropped", peer)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// Tests that if a block is empty (e.g. header only), no body request should be
|
// Tests that if a block is empty (e.g. header only), no body request should be
|
||||||
// made, and instead the header should be assembled into a whole block in itself.
|
// made, and instead the header should be assembled into a whole block in itself.
|
||||||
func TestEmptyShortCircuit68Full(t *testing.T) { testEmptyShortCircuit(t, eth.ETH68, FullSync) }
|
func TestEmptyShortCircuitFull(t *testing.T) { testEmptyShortCircuit(t, eth.ETH69, FullSync) }
|
||||||
func TestEmptyShortCircuit68Snap(t *testing.T) { testEmptyShortCircuit(t, eth.ETH68, SnapSync) }
|
func TestEmptyShortCircuitSnap(t *testing.T) { testEmptyShortCircuit(t, eth.ETH69, SnapSync) }
|
||||||
|
|
||||||
func testEmptyShortCircuit(t *testing.T, protocol uint, mode SyncMode) {
|
func testEmptyShortCircuit(t *testing.T, protocol uint, mode SyncMode) {
|
||||||
success := make(chan struct{})
|
success := make(chan struct{})
|
||||||
|
|
@ -644,8 +605,8 @@ func checkProgress(t *testing.T, d *Downloader, stage string, want ethereum.Sync
|
||||||
|
|
||||||
// Tests that peers below a pre-configured checkpoint block are prevented from
|
// Tests that peers below a pre-configured checkpoint block are prevented from
|
||||||
// being fast-synced from, avoiding potential cheap eclipse attacks.
|
// being fast-synced from, avoiding potential cheap eclipse attacks.
|
||||||
func TestBeaconSync68Full(t *testing.T) { testBeaconSync(t, eth.ETH68, FullSync) }
|
func TestBeaconSyncFull(t *testing.T) { testBeaconSync(t, eth.ETH69, FullSync) }
|
||||||
func TestBeaconSync68Snap(t *testing.T) { testBeaconSync(t, eth.ETH68, SnapSync) }
|
func TestBeaconSyncSnap(t *testing.T) { testBeaconSync(t, eth.ETH69, SnapSync) }
|
||||||
|
|
||||||
func testBeaconSync(t *testing.T, protocol uint, mode SyncMode) {
|
func testBeaconSync(t *testing.T, protocol uint, mode SyncMode) {
|
||||||
var cases = []struct {
|
var cases = []struct {
|
||||||
|
|
@ -690,8 +651,8 @@ func testBeaconSync(t *testing.T, protocol uint, mode SyncMode) {
|
||||||
|
|
||||||
// Tests that synchronisation progress (origin block number, current block number
|
// Tests that synchronisation progress (origin block number, current block number
|
||||||
// and highest block number) is tracked and updated correctly.
|
// and highest block number) is tracked and updated correctly.
|
||||||
func TestSyncProgress68Full(t *testing.T) { testSyncProgress(t, eth.ETH68, FullSync) }
|
func TestSyncProgressFull(t *testing.T) { testSyncProgress(t, eth.ETH69, FullSync) }
|
||||||
func TestSyncProgress68Snap(t *testing.T) { testSyncProgress(t, eth.ETH68, SnapSync) }
|
func TestSyncProgressSnap(t *testing.T) { testSyncProgress(t, eth.ETH69, SnapSync) }
|
||||||
|
|
||||||
func testSyncProgress(t *testing.T, protocol uint, mode SyncMode) {
|
func testSyncProgress(t *testing.T, protocol uint, mode SyncMode) {
|
||||||
success := make(chan struct{})
|
success := make(chan struct{})
|
||||||
|
|
|
||||||
|
|
@ -845,7 +845,7 @@ func TestSkeletonSyncRetrievals(t *testing.T) {
|
||||||
// Create a peer set to feed headers through
|
// Create a peer set to feed headers through
|
||||||
peerset := newPeerSet()
|
peerset := newPeerSet()
|
||||||
for _, peer := range tt.peers {
|
for _, peer := range tt.peers {
|
||||||
peerset.Register(newPeerConnection(peer.id, eth.ETH68, peer, log.New("id", peer.id)))
|
peerset.Register(newPeerConnection(peer.id, eth.ETH69, peer, log.New("id", peer.id)))
|
||||||
}
|
}
|
||||||
// Create a peer dropper to track malicious peers
|
// Create a peer dropper to track malicious peers
|
||||||
dropped := make(map[string]int)
|
dropped := make(map[string]int)
|
||||||
|
|
@ -912,7 +912,7 @@ func TestSkeletonSyncRetrievals(t *testing.T) {
|
||||||
// Apply the post-init events if there's any
|
// Apply the post-init events if there's any
|
||||||
endpeers := tt.peers
|
endpeers := tt.peers
|
||||||
if tt.newPeer != nil {
|
if tt.newPeer != nil {
|
||||||
if err := peerset.Register(newPeerConnection(tt.newPeer.id, eth.ETH68, tt.newPeer, log.New("id", tt.newPeer.id))); err != nil {
|
if err := peerset.Register(newPeerConnection(tt.newPeer.id, eth.ETH69, tt.newPeer, log.New("id", tt.newPeer.id))); err != nil {
|
||||||
t.Errorf("test %d: failed to register new peer: %v", i, err)
|
t.Errorf("test %d: failed to register new peer: %v", i, err)
|
||||||
}
|
}
|
||||||
time.Sleep(time.Millisecond * 50) // given time for peer registration
|
time.Sleep(time.Millisecond * 50) // given time for peer registration
|
||||||
|
|
|
||||||
|
|
@ -24,6 +24,7 @@ var (
|
||||||
txAnnounceInMeter = metrics.NewRegisteredMeter("eth/fetcher/transaction/announces/in", nil)
|
txAnnounceInMeter = metrics.NewRegisteredMeter("eth/fetcher/transaction/announces/in", nil)
|
||||||
txAnnounceKnownMeter = metrics.NewRegisteredMeter("eth/fetcher/transaction/announces/known", nil)
|
txAnnounceKnownMeter = metrics.NewRegisteredMeter("eth/fetcher/transaction/announces/known", nil)
|
||||||
txAnnounceUnderpricedMeter = metrics.NewRegisteredMeter("eth/fetcher/transaction/announces/underpriced", nil)
|
txAnnounceUnderpricedMeter = metrics.NewRegisteredMeter("eth/fetcher/transaction/announces/underpriced", nil)
|
||||||
|
txAnnounceOnchainMeter = metrics.NewRegisteredMeter("eth/fetcher/transaction/announces/onchain", nil)
|
||||||
txAnnounceDOSMeter = metrics.NewRegisteredMeter("eth/fetcher/transaction/announces/dos", nil)
|
txAnnounceDOSMeter = metrics.NewRegisteredMeter("eth/fetcher/transaction/announces/dos", nil)
|
||||||
|
|
||||||
txBroadcastInMeter = metrics.NewRegisteredMeter("eth/fetcher/transaction/broadcasts/in", nil)
|
txBroadcastInMeter = metrics.NewRegisteredMeter("eth/fetcher/transaction/broadcasts/in", nil)
|
||||||
|
|
|
||||||
|
|
@ -27,6 +27,7 @@ import (
|
||||||
"github.com/ethereum/go-ethereum/common"
|
"github.com/ethereum/go-ethereum/common"
|
||||||
"github.com/ethereum/go-ethereum/common/lru"
|
"github.com/ethereum/go-ethereum/common/lru"
|
||||||
"github.com/ethereum/go-ethereum/common/mclock"
|
"github.com/ethereum/go-ethereum/common/mclock"
|
||||||
|
"github.com/ethereum/go-ethereum/core"
|
||||||
"github.com/ethereum/go-ethereum/core/txpool"
|
"github.com/ethereum/go-ethereum/core/txpool"
|
||||||
"github.com/ethereum/go-ethereum/core/types"
|
"github.com/ethereum/go-ethereum/core/types"
|
||||||
"github.com/ethereum/go-ethereum/log"
|
"github.com/ethereum/go-ethereum/log"
|
||||||
|
|
@ -71,6 +72,11 @@ const (
|
||||||
|
|
||||||
// addTxsBatchSize it the max number of transactions to add in a single batch from a peer.
|
// addTxsBatchSize it the max number of transactions to add in a single batch from a peer.
|
||||||
addTxsBatchSize = 128
|
addTxsBatchSize = 128
|
||||||
|
|
||||||
|
// txOnChainCacheLimit is number of on-chain transactions to keep in a cache to avoid
|
||||||
|
// re-fetching them soon after they are mined.
|
||||||
|
// Approx 1MB for 30 minutes of transactions at 18 tps
|
||||||
|
txOnChainCacheLimit = 32768
|
||||||
)
|
)
|
||||||
|
|
||||||
var (
|
var (
|
||||||
|
|
@ -152,6 +158,9 @@ type TxFetcher struct {
|
||||||
txSeq uint64 // Unique transaction sequence number
|
txSeq uint64 // Unique transaction sequence number
|
||||||
underpriced *lru.Cache[common.Hash, time.Time] // Transactions discarded as too cheap (don't re-fetch)
|
underpriced *lru.Cache[common.Hash, time.Time] // Transactions discarded as too cheap (don't re-fetch)
|
||||||
|
|
||||||
|
chain *core.BlockChain // Blockchain interface for on-chain checks
|
||||||
|
txOnChainCache *lru.Cache[common.Hash, struct{}] // Cache to avoid fetching once the tx gets on chain
|
||||||
|
|
||||||
// Stage 1: Waiting lists for newly discovered transactions that might be
|
// Stage 1: Waiting lists for newly discovered transactions that might be
|
||||||
// broadcast without needing explicit request/reply round trips.
|
// broadcast without needing explicit request/reply round trips.
|
||||||
waitlist map[common.Hash]map[string]struct{} // Transactions waiting for an potential broadcast
|
waitlist map[common.Hash]map[string]struct{} // Transactions waiting for an potential broadcast
|
||||||
|
|
@ -184,36 +193,40 @@ type TxFetcher struct {
|
||||||
|
|
||||||
// NewTxFetcher creates a transaction fetcher to retrieve transaction
|
// NewTxFetcher creates a transaction fetcher to retrieve transaction
|
||||||
// based on hash announcements.
|
// based on hash announcements.
|
||||||
func NewTxFetcher(validateMeta func(common.Hash, byte) error, addTxs func([]*types.Transaction) []error, fetchTxs func(string, []common.Hash) error, dropPeer func(string)) *TxFetcher {
|
// Chain can be nil to disable on-chain checks.
|
||||||
return NewTxFetcherForTests(validateMeta, addTxs, fetchTxs, dropPeer, mclock.System{}, time.Now, nil)
|
func NewTxFetcher(chain *core.BlockChain, validateMeta func(common.Hash, byte) error, addTxs func([]*types.Transaction) []error, fetchTxs func(string, []common.Hash) error, dropPeer func(string)) *TxFetcher {
|
||||||
|
return NewTxFetcherForTests(chain, validateMeta, addTxs, fetchTxs, dropPeer, mclock.System{}, time.Now, nil)
|
||||||
}
|
}
|
||||||
|
|
||||||
// NewTxFetcherForTests is a testing method to mock out the realtime clock with
|
// NewTxFetcherForTests is a testing method to mock out the realtime clock with
|
||||||
// a simulated version and the internal randomness with a deterministic one.
|
// a simulated version and the internal randomness with a deterministic one.
|
||||||
|
// Chain can be nil to disable on-chain checks.
|
||||||
func NewTxFetcherForTests(
|
func NewTxFetcherForTests(
|
||||||
validateMeta func(common.Hash, byte) error, addTxs func([]*types.Transaction) []error, fetchTxs func(string, []common.Hash) error, dropPeer func(string),
|
chain *core.BlockChain, validateMeta func(common.Hash, byte) error, addTxs func([]*types.Transaction) []error, fetchTxs func(string, []common.Hash) error, dropPeer func(string),
|
||||||
clock mclock.Clock, realTime func() time.Time, rand *mrand.Rand) *TxFetcher {
|
clock mclock.Clock, realTime func() time.Time, rand *mrand.Rand) *TxFetcher {
|
||||||
return &TxFetcher{
|
return &TxFetcher{
|
||||||
notify: make(chan *txAnnounce),
|
notify: make(chan *txAnnounce),
|
||||||
cleanup: make(chan *txDelivery),
|
cleanup: make(chan *txDelivery),
|
||||||
drop: make(chan *txDrop),
|
drop: make(chan *txDrop),
|
||||||
quit: make(chan struct{}),
|
quit: make(chan struct{}),
|
||||||
waitlist: make(map[common.Hash]map[string]struct{}),
|
waitlist: make(map[common.Hash]map[string]struct{}),
|
||||||
waittime: make(map[common.Hash]mclock.AbsTime),
|
waittime: make(map[common.Hash]mclock.AbsTime),
|
||||||
waitslots: make(map[string]map[common.Hash]*txMetadataWithSeq),
|
waitslots: make(map[string]map[common.Hash]*txMetadataWithSeq),
|
||||||
announces: make(map[string]map[common.Hash]*txMetadataWithSeq),
|
announces: make(map[string]map[common.Hash]*txMetadataWithSeq),
|
||||||
announced: make(map[common.Hash]map[string]struct{}),
|
announced: make(map[common.Hash]map[string]struct{}),
|
||||||
fetching: make(map[common.Hash]string),
|
fetching: make(map[common.Hash]string),
|
||||||
requests: make(map[string]*txRequest),
|
requests: make(map[string]*txRequest),
|
||||||
alternates: make(map[common.Hash]map[string]struct{}),
|
alternates: make(map[common.Hash]map[string]struct{}),
|
||||||
underpriced: lru.NewCache[common.Hash, time.Time](maxTxUnderpricedSetSize),
|
underpriced: lru.NewCache[common.Hash, time.Time](maxTxUnderpricedSetSize),
|
||||||
validateMeta: validateMeta,
|
txOnChainCache: lru.NewCache[common.Hash, struct{}](txOnChainCacheLimit),
|
||||||
addTxs: addTxs,
|
chain: chain,
|
||||||
fetchTxs: fetchTxs,
|
validateMeta: validateMeta,
|
||||||
dropPeer: dropPeer,
|
addTxs: addTxs,
|
||||||
clock: clock,
|
fetchTxs: fetchTxs,
|
||||||
realTime: realTime,
|
dropPeer: dropPeer,
|
||||||
rand: rand,
|
clock: clock,
|
||||||
|
realTime: realTime,
|
||||||
|
rand: rand,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
@ -233,6 +246,7 @@ func (f *TxFetcher) Notify(peer string, types []byte, sizes []uint32, hashes []c
|
||||||
unknownMetas = make([]txMetadata, 0, len(hashes))
|
unknownMetas = make([]txMetadata, 0, len(hashes))
|
||||||
|
|
||||||
duplicate int64
|
duplicate int64
|
||||||
|
onchain int64
|
||||||
underpriced int64
|
underpriced int64
|
||||||
)
|
)
|
||||||
for i, hash := range hashes {
|
for i, hash := range hashes {
|
||||||
|
|
@ -245,6 +259,12 @@ func (f *TxFetcher) Notify(peer string, types []byte, sizes []uint32, hashes []c
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// check on chain as well (no need to check limbo separately, as chain checks limbo too)
|
||||||
|
if _, exist := f.txOnChainCache.Get(hash); exist {
|
||||||
|
onchain++
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
|
||||||
if f.isKnownUnderpriced(hash) {
|
if f.isKnownUnderpriced(hash) {
|
||||||
underpriced++
|
underpriced++
|
||||||
continue
|
continue
|
||||||
|
|
@ -259,6 +279,7 @@ func (f *TxFetcher) Notify(peer string, types []byte, sizes []uint32, hashes []c
|
||||||
}
|
}
|
||||||
txAnnounceKnownMeter.Mark(duplicate)
|
txAnnounceKnownMeter.Mark(duplicate)
|
||||||
txAnnounceUnderpricedMeter.Mark(underpriced)
|
txAnnounceUnderpricedMeter.Mark(underpriced)
|
||||||
|
txAnnounceOnchainMeter.Mark(onchain)
|
||||||
|
|
||||||
// If anything's left to announce, push it into the internal loop
|
// If anything's left to announce, push it into the internal loop
|
||||||
if len(unknownHashes) == 0 {
|
if len(unknownHashes) == 0 {
|
||||||
|
|
@ -412,7 +433,18 @@ func (f *TxFetcher) loop() {
|
||||||
|
|
||||||
waitTrigger = make(chan struct{}, 1)
|
waitTrigger = make(chan struct{}, 1)
|
||||||
timeoutTrigger = make(chan struct{}, 1)
|
timeoutTrigger = make(chan struct{}, 1)
|
||||||
|
|
||||||
|
oldHead *types.Header
|
||||||
)
|
)
|
||||||
|
|
||||||
|
// Subscribe to chain events to know when transactions are added to chain
|
||||||
|
var headEventCh chan core.ChainEvent
|
||||||
|
if f.chain != nil {
|
||||||
|
headEventCh = make(chan core.ChainEvent, 10)
|
||||||
|
sub := f.chain.SubscribeChainEvent(headEventCh)
|
||||||
|
defer sub.Unsubscribe()
|
||||||
|
}
|
||||||
|
|
||||||
for {
|
for {
|
||||||
select {
|
select {
|
||||||
case ann := <-f.notify:
|
case ann := <-f.notify:
|
||||||
|
|
@ -837,6 +869,21 @@ func (f *TxFetcher) loop() {
|
||||||
f.rescheduleTimeout(timeoutTimer, timeoutTrigger)
|
f.rescheduleTimeout(timeoutTimer, timeoutTrigger)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
case ev := <-headEventCh:
|
||||||
|
// New head(s) added
|
||||||
|
newHead := ev.Header
|
||||||
|
if oldHead != nil && newHead.ParentHash != oldHead.Hash() {
|
||||||
|
// Reorg or setHead detected, clear the cache. We could be smarter here and
|
||||||
|
// only remove/add the diff, but this is simpler and not being exact here
|
||||||
|
// only results in a few more fetches.
|
||||||
|
f.txOnChainCache.Purge()
|
||||||
|
}
|
||||||
|
oldHead = newHead
|
||||||
|
// Add all transactions from the new block to the on-chain cache
|
||||||
|
for _, tx := range ev.Transactions {
|
||||||
|
f.txOnChainCache.Add(tx.Hash(), struct{}{})
|
||||||
|
}
|
||||||
|
|
||||||
case <-f.quit:
|
case <-f.quit:
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
|
||||||
|
|
@ -91,6 +91,7 @@ type txFetcherTest struct {
|
||||||
// and deterministic randomness.
|
// and deterministic randomness.
|
||||||
func newTestTxFetcher() *TxFetcher {
|
func newTestTxFetcher() *TxFetcher {
|
||||||
return NewTxFetcher(
|
return NewTxFetcher(
|
||||||
|
nil,
|
||||||
func(common.Hash, byte) error { return nil },
|
func(common.Hash, byte) error { return nil },
|
||||||
func(txs []*types.Transaction) []error {
|
func(txs []*types.Transaction) []error {
|
||||||
return make([]error, len(txs))
|
return make([]error, len(txs))
|
||||||
|
|
@ -2191,6 +2192,7 @@ func TestTransactionForgotten(t *testing.T) {
|
||||||
}
|
}
|
||||||
|
|
||||||
fetcher := NewTxFetcherForTests(
|
fetcher := NewTxFetcherForTests(
|
||||||
|
nil,
|
||||||
func(common.Hash, byte) error { return nil },
|
func(common.Hash, byte) error { return nil },
|
||||||
func(txs []*types.Transaction) []error {
|
func(txs []*types.Transaction) []error {
|
||||||
errs := make([]error, len(txs))
|
errs := make([]error, len(txs))
|
||||||
|
|
|
||||||
|
|
@ -20,7 +20,6 @@ import (
|
||||||
"context"
|
"context"
|
||||||
"errors"
|
"errors"
|
||||||
"fmt"
|
"fmt"
|
||||||
"math"
|
|
||||||
"math/big"
|
"math/big"
|
||||||
|
|
||||||
"github.com/ethereum/go-ethereum/common"
|
"github.com/ethereum/go-ethereum/common"
|
||||||
|
|
@ -268,7 +267,7 @@ func run(ctx context.Context, call *core.Message, opts *Options) (*core.Executio
|
||||||
evm.Cancel()
|
evm.Cancel()
|
||||||
}()
|
}()
|
||||||
// Execute the call, returning a wrapped error or the result
|
// Execute the call, returning a wrapped error or the result
|
||||||
result, err := core.ApplyMessage(evm, call, new(core.GasPool).AddGas(math.MaxUint64))
|
result, err := core.ApplyMessage(evm, call, nil)
|
||||||
if vmerr := dirtyState.Error(); vmerr != nil {
|
if vmerr := dirtyState.Error(); vmerr != nil {
|
||||||
return nil, vmerr
|
return nil, vmerr
|
||||||
}
|
}
|
||||||
|
|
|
||||||
Some files were not shown because too many files have changed in this diff Show more
Loading…
Reference in a new issue