forked from forks/go-ethereum
The total difficulty is the sum of all block difficulties from genesis to a certain block. This value was used in PoW for deciding which chain is heavier, and thus which chain to select. Since PoS has a different fork selection algorithm, all blocks since the merge have a difficulty of 0, and all total difficulties are the same for the past 2 years. Whilst the TDs are mostly useless nowadays, there was never really a reason to mess around removing them since they are so tiny. This reasoning changes when we go down the path of pruned chain history. In order to reconstruct any TD, we **must** retrieve all the headers from chain head to genesis and then iterate all the difficulties to compute the TD. In a world where we completely prune past chain segments (bodies, receipts, headers), it is not possible to reconstruct the TD at all. In a world where we still keep chain headers and prune only the rest, reconstructing it possible as long as we process (or download) the chain forward from genesis, but trying to snap sync the head first and backfill later hits the same issue, the TD becomes impossible to calculate until genesis is backfilled. All in all, the TD is a messy out-of-state, out-of-consensus computed field that is overall useless nowadays, but code relying on it forces the client into certain modes of operation and prevents other modes or other optimizations. This PR completely nukes out the TD from the node. It doesn't compute it, it doesn't operate on it, it's as if it didn't even exist. Caveats: - Whenever we have APIs that return TD (devp2p handshake, tracer, etc.) we return a TD of 0. - For era files, we recompute the TD during export time (fairly quick) to retain the format content. - It is not possible to "verify" the merge point (i.e. with TD gone, TTD is useless). Since we're not verifying PoW any more, just blindly trust it, not verifying but blindly trusting the many year old merge point seems just the same trust model. - Our tests still need to be able to generate pre and post merge blocks, so they need a new way to split the merge without TTD. The PR introduces a settable ttdBlock field on the consensus object which is used by tests as the block where originally the TTD happened. This is not needed for live nodes, we never want to generate old blocks. - One merge transition consensus test was disabled. With a non-operational TD, testing how the client reacts to TTD is useless, it cannot react. Questions: - Should we also drop total terminal difficulty from the genesis json? It's a number we cannot react on any more, so maybe it would be cleaner to get rid of even more concepts. --------- Co-authored-by: Gary Rong <garyrong0905@gmail.com>
827 lines
24 KiB
Go
827 lines
24 KiB
Go
// Copyright 2020 The go-ethereum Authors
|
|
// This file is part of go-ethereum.
|
|
//
|
|
// go-ethereum is free software: you can redistribute it and/or modify
|
|
// it under the terms of the GNU General Public License as published by
|
|
// the Free Software Foundation, either version 3 of the License, or
|
|
// (at your option) any later version.
|
|
//
|
|
// go-ethereum is distributed in the hope that it will be useful,
|
|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
// GNU General Public License for more details.
|
|
//
|
|
// You should have received a copy of the GNU General Public License
|
|
// along with go-ethereum. If not, see <http://www.gnu.org/licenses/>.
|
|
|
|
package ethtest
|
|
|
|
import (
|
|
"crypto/rand"
|
|
"reflect"
|
|
|
|
"github.com/ethereum/go-ethereum/common"
|
|
"github.com/ethereum/go-ethereum/consensus/misc/eip4844"
|
|
"github.com/ethereum/go-ethereum/core/types"
|
|
"github.com/ethereum/go-ethereum/crypto"
|
|
"github.com/ethereum/go-ethereum/crypto/kzg4844"
|
|
"github.com/ethereum/go-ethereum/eth/protocols/eth"
|
|
"github.com/ethereum/go-ethereum/internal/utesting"
|
|
"github.com/ethereum/go-ethereum/p2p"
|
|
"github.com/ethereum/go-ethereum/p2p/enode"
|
|
"github.com/holiman/uint256"
|
|
)
|
|
|
|
// Suite represents a structure used to test a node's conformance
|
|
// to the eth protocol.
|
|
type Suite struct {
|
|
Dest *enode.Node
|
|
chain *Chain
|
|
engine *EngineClient
|
|
}
|
|
|
|
// NewSuite creates and returns a new eth-test suite that can
|
|
// be used to test the given node against the given blockchain
|
|
// data.
|
|
func NewSuite(dest *enode.Node, chainDir, engineURL, jwt string) (*Suite, error) {
|
|
chain, err := NewChain(chainDir)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
engine, err := NewEngineClient(chainDir, engineURL, jwt)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
return &Suite{
|
|
Dest: dest,
|
|
chain: chain,
|
|
engine: engine,
|
|
}, nil
|
|
}
|
|
|
|
func (s *Suite) EthTests() []utesting.Test {
|
|
return []utesting.Test{
|
|
// status
|
|
{Name: "Status", Fn: s.TestStatus},
|
|
// get block headers
|
|
{Name: "GetBlockHeaders", Fn: s.TestGetBlockHeaders},
|
|
{Name: "SimultaneousRequests", Fn: s.TestSimultaneousRequests},
|
|
{Name: "SameRequestID", Fn: s.TestSameRequestID},
|
|
{Name: "ZeroRequestID", Fn: s.TestZeroRequestID},
|
|
// get block bodies
|
|
{Name: "GetBlockBodies", Fn: s.TestGetBlockBodies},
|
|
// // malicious handshakes + status
|
|
{Name: "MaliciousHandshake", Fn: s.TestMaliciousHandshake},
|
|
// test transactions
|
|
{Name: "LargeTxRequest", Fn: s.TestLargeTxRequest, Slow: true},
|
|
{Name: "Transaction", Fn: s.TestTransaction},
|
|
{Name: "InvalidTxs", Fn: s.TestInvalidTxs},
|
|
{Name: "NewPooledTxs", Fn: s.TestNewPooledTxs},
|
|
{Name: "BlobViolations", Fn: s.TestBlobViolations},
|
|
}
|
|
}
|
|
|
|
func (s *Suite) SnapTests() []utesting.Test {
|
|
return []utesting.Test{
|
|
{Name: "Status", Fn: s.TestSnapStatus},
|
|
{Name: "AccountRange", Fn: s.TestSnapGetAccountRange},
|
|
{Name: "GetByteCodes", Fn: s.TestSnapGetByteCodes},
|
|
{Name: "GetTrieNodes", Fn: s.TestSnapTrieNodes},
|
|
{Name: "GetStorageRanges", Fn: s.TestSnapGetStorageRanges},
|
|
}
|
|
}
|
|
|
|
func (s *Suite) TestStatus(t *utesting.T) {
|
|
t.Log(`This test is just a sanity check. It performs an eth protocol handshake.`)
|
|
|
|
conn, err := s.dial()
|
|
if err != nil {
|
|
t.Fatalf("dial failed: %v", err)
|
|
}
|
|
defer conn.Close()
|
|
if err := conn.peer(s.chain, nil); err != nil {
|
|
t.Fatalf("peering failed: %v", err)
|
|
}
|
|
}
|
|
|
|
// headersMatch returns whether the received headers match the given request
|
|
func headersMatch(expected []*types.Header, headers []*types.Header) bool {
|
|
return reflect.DeepEqual(expected, headers)
|
|
}
|
|
|
|
func (s *Suite) TestGetBlockHeaders(t *utesting.T) {
|
|
t.Log(`This test requests block headers from the node.`)
|
|
|
|
conn, err := s.dial()
|
|
if err != nil {
|
|
t.Fatalf("dial failed: %v", err)
|
|
}
|
|
defer conn.Close()
|
|
if err = conn.peer(s.chain, nil); err != nil {
|
|
t.Fatalf("peering failed: %v", err)
|
|
}
|
|
// Send headers request.
|
|
req := ð.GetBlockHeadersPacket{
|
|
RequestId: 33,
|
|
GetBlockHeadersRequest: ð.GetBlockHeadersRequest{
|
|
Origin: eth.HashOrNumber{Hash: s.chain.blocks[1].Hash()},
|
|
Amount: 2,
|
|
Skip: 1,
|
|
Reverse: false,
|
|
},
|
|
}
|
|
// Read headers response.
|
|
if err := conn.Write(ethProto, eth.GetBlockHeadersMsg, req); err != nil {
|
|
t.Fatalf("could not write to connection: %v", err)
|
|
}
|
|
headers := new(eth.BlockHeadersPacket)
|
|
if err := conn.ReadMsg(ethProto, eth.BlockHeadersMsg, &headers); err != nil {
|
|
t.Fatalf("error reading msg: %v", err)
|
|
}
|
|
if got, want := headers.RequestId, req.RequestId; got != want {
|
|
t.Fatalf("unexpected request id")
|
|
}
|
|
// Check for correct headers.
|
|
expected, err := s.chain.GetHeaders(req)
|
|
if err != nil {
|
|
t.Fatalf("failed to get headers for given request: %v", err)
|
|
}
|
|
if !headersMatch(expected, headers.BlockHeadersRequest) {
|
|
t.Fatalf("header mismatch: \nexpected %v \ngot %v", expected, headers)
|
|
}
|
|
}
|
|
|
|
func (s *Suite) TestSimultaneousRequests(t *utesting.T) {
|
|
t.Log(`This test requests blocks headers from the node, performing two requests
|
|
concurrently, with different request IDs.`)
|
|
|
|
conn, err := s.dial()
|
|
if err != nil {
|
|
t.Fatalf("dial failed: %v", err)
|
|
}
|
|
defer conn.Close()
|
|
if err := conn.peer(s.chain, nil); err != nil {
|
|
t.Fatalf("peering failed: %v", err)
|
|
}
|
|
|
|
// Create two different requests.
|
|
req1 := ð.GetBlockHeadersPacket{
|
|
RequestId: uint64(111),
|
|
GetBlockHeadersRequest: ð.GetBlockHeadersRequest{
|
|
Origin: eth.HashOrNumber{
|
|
Hash: s.chain.blocks[1].Hash(),
|
|
},
|
|
Amount: 2,
|
|
Skip: 1,
|
|
Reverse: false,
|
|
},
|
|
}
|
|
req2 := ð.GetBlockHeadersPacket{
|
|
RequestId: uint64(222),
|
|
GetBlockHeadersRequest: ð.GetBlockHeadersRequest{
|
|
Origin: eth.HashOrNumber{
|
|
Hash: s.chain.blocks[1].Hash(),
|
|
},
|
|
Amount: 4,
|
|
Skip: 1,
|
|
Reverse: false,
|
|
},
|
|
}
|
|
|
|
// Send both requests.
|
|
if err := conn.Write(ethProto, eth.GetBlockHeadersMsg, req1); err != nil {
|
|
t.Fatalf("failed to write to connection: %v", err)
|
|
}
|
|
if err := conn.Write(ethProto, eth.GetBlockHeadersMsg, req2); err != nil {
|
|
t.Fatalf("failed to write to connection: %v", err)
|
|
}
|
|
|
|
// Wait for responses.
|
|
headers1 := new(eth.BlockHeadersPacket)
|
|
if err := conn.ReadMsg(ethProto, eth.BlockHeadersMsg, &headers1); err != nil {
|
|
t.Fatalf("error reading block headers msg: %v", err)
|
|
}
|
|
if got, want := headers1.RequestId, req1.RequestId; got != want {
|
|
t.Fatalf("unexpected request id in response: got %d, want %d", got, want)
|
|
}
|
|
headers2 := new(eth.BlockHeadersPacket)
|
|
if err := conn.ReadMsg(ethProto, eth.BlockHeadersMsg, &headers2); err != nil {
|
|
t.Fatalf("error reading block headers msg: %v", err)
|
|
}
|
|
if got, want := headers2.RequestId, req2.RequestId; got != want {
|
|
t.Fatalf("unexpected request id in response: got %d, want %d", got, want)
|
|
}
|
|
|
|
// Check received headers for accuracy.
|
|
if expected, err := s.chain.GetHeaders(req1); err != nil {
|
|
t.Fatalf("failed to get expected headers for request 1: %v", err)
|
|
} else if !headersMatch(expected, headers1.BlockHeadersRequest) {
|
|
t.Fatalf("header mismatch: \nexpected %v \ngot %v", expected, headers1)
|
|
}
|
|
if expected, err := s.chain.GetHeaders(req2); err != nil {
|
|
t.Fatalf("failed to get expected headers for request 2: %v", err)
|
|
} else if !headersMatch(expected, headers2.BlockHeadersRequest) {
|
|
t.Fatalf("header mismatch: \nexpected %v \ngot %v", expected, headers2)
|
|
}
|
|
}
|
|
|
|
func (s *Suite) TestSameRequestID(t *utesting.T) {
|
|
t.Log(`This test requests block headers, performing two concurrent requests with the
|
|
same request ID. The node should handle the request by responding to both requests.`)
|
|
|
|
conn, err := s.dial()
|
|
if err != nil {
|
|
t.Fatalf("dial failed: %v", err)
|
|
}
|
|
defer conn.Close()
|
|
if err := conn.peer(s.chain, nil); err != nil {
|
|
t.Fatalf("peering failed: %v", err)
|
|
}
|
|
|
|
// Create two different requests with the same ID.
|
|
reqID := uint64(1234)
|
|
request1 := ð.GetBlockHeadersPacket{
|
|
RequestId: reqID,
|
|
GetBlockHeadersRequest: ð.GetBlockHeadersRequest{
|
|
Origin: eth.HashOrNumber{
|
|
Number: 1,
|
|
},
|
|
Amount: 2,
|
|
},
|
|
}
|
|
request2 := ð.GetBlockHeadersPacket{
|
|
RequestId: reqID,
|
|
GetBlockHeadersRequest: ð.GetBlockHeadersRequest{
|
|
Origin: eth.HashOrNumber{
|
|
Number: 33,
|
|
},
|
|
Amount: 2,
|
|
},
|
|
}
|
|
|
|
// Send the requests.
|
|
if err = conn.Write(ethProto, eth.GetBlockHeadersMsg, request1); err != nil {
|
|
t.Fatalf("failed to write to connection: %v", err)
|
|
}
|
|
if err = conn.Write(ethProto, eth.GetBlockHeadersMsg, request2); err != nil {
|
|
t.Fatalf("failed to write to connection: %v", err)
|
|
}
|
|
|
|
// Wait for the responses.
|
|
headers1 := new(eth.BlockHeadersPacket)
|
|
if err := conn.ReadMsg(ethProto, eth.BlockHeadersMsg, &headers1); err != nil {
|
|
t.Fatalf("error reading from connection: %v", err)
|
|
}
|
|
if got, want := headers1.RequestId, request1.RequestId; got != want {
|
|
t.Fatalf("unexpected request id: got %d, want %d", got, want)
|
|
}
|
|
headers2 := new(eth.BlockHeadersPacket)
|
|
if err := conn.ReadMsg(ethProto, eth.BlockHeadersMsg, &headers2); err != nil {
|
|
t.Fatalf("error reading from connection: %v", err)
|
|
}
|
|
if got, want := headers2.RequestId, request2.RequestId; got != want {
|
|
t.Fatalf("unexpected request id: got %d, want %d", got, want)
|
|
}
|
|
|
|
// Check if headers match.
|
|
if expected, err := s.chain.GetHeaders(request1); err != nil {
|
|
t.Fatalf("failed to get expected block headers: %v", err)
|
|
} else if !headersMatch(expected, headers1.BlockHeadersRequest) {
|
|
t.Fatalf("header mismatch: \nexpected %v \ngot %v", expected, headers1)
|
|
}
|
|
if expected, err := s.chain.GetHeaders(request2); err != nil {
|
|
t.Fatalf("failed to get expected block headers: %v", err)
|
|
} else if !headersMatch(expected, headers2.BlockHeadersRequest) {
|
|
t.Fatalf("header mismatch: \nexpected %v \ngot %v", expected, headers2)
|
|
}
|
|
}
|
|
|
|
func (s *Suite) TestZeroRequestID(t *utesting.T) {
|
|
t.Log(`This test sends a GetBlockHeaders message with a request-id of zero,
|
|
and expects a response.`)
|
|
|
|
conn, err := s.dial()
|
|
if err != nil {
|
|
t.Fatalf("dial failed: %v", err)
|
|
}
|
|
defer conn.Close()
|
|
if err := conn.peer(s.chain, nil); err != nil {
|
|
t.Fatalf("peering failed: %v", err)
|
|
}
|
|
req := ð.GetBlockHeadersPacket{
|
|
GetBlockHeadersRequest: ð.GetBlockHeadersRequest{
|
|
Origin: eth.HashOrNumber{Number: 0},
|
|
Amount: 2,
|
|
},
|
|
}
|
|
// Read headers response.
|
|
if err := conn.Write(ethProto, eth.GetBlockHeadersMsg, req); err != nil {
|
|
t.Fatalf("could not write to connection: %v", err)
|
|
}
|
|
headers := new(eth.BlockHeadersPacket)
|
|
if err := conn.ReadMsg(ethProto, eth.BlockHeadersMsg, &headers); err != nil {
|
|
t.Fatalf("error reading msg: %v", err)
|
|
}
|
|
if got, want := headers.RequestId, req.RequestId; got != want {
|
|
t.Fatalf("unexpected request id")
|
|
}
|
|
if expected, err := s.chain.GetHeaders(req); err != nil {
|
|
t.Fatalf("failed to get expected block headers: %v", err)
|
|
} else if !headersMatch(expected, headers.BlockHeadersRequest) {
|
|
t.Fatalf("header mismatch: \nexpected %v \ngot %v", expected, headers)
|
|
}
|
|
}
|
|
|
|
func (s *Suite) TestGetBlockBodies(t *utesting.T) {
|
|
t.Log(`This test sends GetBlockBodies requests to the node for known blocks in the test chain.`)
|
|
|
|
conn, err := s.dial()
|
|
if err != nil {
|
|
t.Fatalf("dial failed: %v", err)
|
|
}
|
|
defer conn.Close()
|
|
if err := conn.peer(s.chain, nil); err != nil {
|
|
t.Fatalf("peering failed: %v", err)
|
|
}
|
|
// Create block bodies request.
|
|
req := ð.GetBlockBodiesPacket{
|
|
RequestId: 55,
|
|
GetBlockBodiesRequest: eth.GetBlockBodiesRequest{
|
|
s.chain.blocks[54].Hash(),
|
|
s.chain.blocks[75].Hash(),
|
|
},
|
|
}
|
|
if err := conn.Write(ethProto, eth.GetBlockBodiesMsg, req); err != nil {
|
|
t.Fatalf("could not write to connection: %v", err)
|
|
}
|
|
// Wait for response.
|
|
resp := new(eth.BlockBodiesPacket)
|
|
if err := conn.ReadMsg(ethProto, eth.BlockBodiesMsg, &resp); err != nil {
|
|
t.Fatalf("error reading block bodies msg: %v", err)
|
|
}
|
|
if got, want := resp.RequestId, req.RequestId; got != want {
|
|
t.Fatalf("unexpected request id in respond", got, want)
|
|
}
|
|
bodies := resp.BlockBodiesResponse
|
|
if len(bodies) != len(req.GetBlockBodiesRequest) {
|
|
t.Fatalf("wrong bodies in response: expected %d bodies, got %d", len(req.GetBlockBodiesRequest), len(bodies))
|
|
}
|
|
}
|
|
|
|
// randBuf makes a random buffer size kilobytes large.
|
|
func randBuf(size int) []byte {
|
|
buf := make([]byte, size*1024)
|
|
rand.Read(buf)
|
|
return buf
|
|
}
|
|
|
|
func (s *Suite) TestMaliciousHandshake(t *utesting.T) {
|
|
t.Log(`This test tries to send malicious data during the devp2p handshake, in various ways.`)
|
|
|
|
// Write hello to client.
|
|
var (
|
|
key, _ = crypto.GenerateKey()
|
|
pub0 = crypto.FromECDSAPub(&key.PublicKey)[1:]
|
|
version = eth.ProtocolVersions[0]
|
|
)
|
|
handshakes := []*protoHandshake{
|
|
{
|
|
Version: 5,
|
|
Caps: []p2p.Cap{
|
|
{Name: string(randBuf(2)), Version: version},
|
|
},
|
|
ID: pub0,
|
|
},
|
|
{
|
|
Version: 5,
|
|
Caps: []p2p.Cap{
|
|
{Name: "eth", Version: version},
|
|
},
|
|
ID: append(pub0, byte(0)),
|
|
},
|
|
{
|
|
Version: 5,
|
|
Caps: []p2p.Cap{
|
|
{Name: "eth", Version: version},
|
|
},
|
|
ID: append(pub0, pub0...),
|
|
},
|
|
{
|
|
Version: 5,
|
|
Caps: []p2p.Cap{
|
|
{Name: "eth", Version: version},
|
|
},
|
|
ID: randBuf(2),
|
|
},
|
|
{
|
|
Version: 5,
|
|
Caps: []p2p.Cap{
|
|
{Name: string(randBuf(2)), Version: version},
|
|
},
|
|
ID: randBuf(2),
|
|
},
|
|
}
|
|
for _, handshake := range handshakes {
|
|
conn, err := s.dialAs(key)
|
|
if err != nil {
|
|
t.Fatalf("dial failed: %v", err)
|
|
}
|
|
defer conn.Close()
|
|
|
|
if err := conn.Write(ethProto, handshakeMsg, handshake); err != nil {
|
|
t.Fatalf("could not write to connection: %v", err)
|
|
}
|
|
// Check that the peer disconnected
|
|
for i := 0; i < 2; i++ {
|
|
code, _, err := conn.Read()
|
|
if err != nil {
|
|
// Client may have disconnected without sending disconnect msg.
|
|
continue
|
|
}
|
|
switch code {
|
|
case discMsg:
|
|
case handshakeMsg:
|
|
// Discard one hello as Hello's are sent concurrently
|
|
continue
|
|
default:
|
|
t.Fatalf("unexpected msg: code %d", code)
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
func (s *Suite) TestTransaction(t *utesting.T) {
|
|
t.Log(`This test sends a valid transaction to the node and checks if the
|
|
transaction gets propagated.`)
|
|
|
|
// Nudge client out of syncing mode to accept pending txs.
|
|
if err := s.engine.sendForkchoiceUpdated(); err != nil {
|
|
t.Fatalf("failed to send next block: %v", err)
|
|
}
|
|
from, nonce := s.chain.GetSender(0)
|
|
inner := &types.DynamicFeeTx{
|
|
ChainID: s.chain.config.ChainID,
|
|
Nonce: nonce,
|
|
GasTipCap: common.Big1,
|
|
GasFeeCap: s.chain.Head().BaseFee(),
|
|
Gas: 30000,
|
|
To: &common.Address{0xaa},
|
|
Value: common.Big1,
|
|
}
|
|
tx, err := s.chain.SignTx(from, types.NewTx(inner))
|
|
if err != nil {
|
|
t.Fatalf("failed to sign tx: %v", err)
|
|
}
|
|
if err := s.sendTxs(t, []*types.Transaction{tx}); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
s.chain.IncNonce(from, 1)
|
|
}
|
|
|
|
func (s *Suite) TestInvalidTxs(t *utesting.T) {
|
|
t.Log(`This test sends several kinds of invalid transactions and checks that the node
|
|
does not propagate them.`)
|
|
|
|
// Nudge client out of syncing mode to accept pending txs.
|
|
if err := s.engine.sendForkchoiceUpdated(); err != nil {
|
|
t.Fatalf("failed to send next block: %v", err)
|
|
}
|
|
|
|
from, nonce := s.chain.GetSender(0)
|
|
inner := &types.DynamicFeeTx{
|
|
ChainID: s.chain.config.ChainID,
|
|
Nonce: nonce,
|
|
GasTipCap: common.Big1,
|
|
GasFeeCap: s.chain.Head().BaseFee(),
|
|
Gas: 30000,
|
|
To: &common.Address{0xaa},
|
|
}
|
|
tx, err := s.chain.SignTx(from, types.NewTx(inner))
|
|
if err != nil {
|
|
t.Fatalf("failed to sign tx: %v", err)
|
|
}
|
|
if err := s.sendTxs(t, []*types.Transaction{tx}); err != nil {
|
|
t.Fatalf("failed to send txs: %v", err)
|
|
}
|
|
s.chain.IncNonce(from, 1)
|
|
|
|
inners := []*types.DynamicFeeTx{
|
|
// Nonce already used
|
|
{
|
|
ChainID: s.chain.config.ChainID,
|
|
Nonce: nonce - 1,
|
|
GasTipCap: common.Big1,
|
|
GasFeeCap: s.chain.Head().BaseFee(),
|
|
Gas: 100000,
|
|
},
|
|
// Value exceeds balance
|
|
{
|
|
Nonce: nonce,
|
|
GasTipCap: common.Big1,
|
|
GasFeeCap: s.chain.Head().BaseFee(),
|
|
Gas: 100000,
|
|
Value: s.chain.Balance(from),
|
|
},
|
|
// Gas limit too low
|
|
{
|
|
Nonce: nonce,
|
|
GasTipCap: common.Big1,
|
|
GasFeeCap: s.chain.Head().BaseFee(),
|
|
Gas: 1337,
|
|
},
|
|
// Code size too large
|
|
{
|
|
Nonce: nonce,
|
|
GasTipCap: common.Big1,
|
|
GasFeeCap: s.chain.Head().BaseFee(),
|
|
Data: randBuf(50),
|
|
Gas: 1_000_000,
|
|
},
|
|
// Data too large
|
|
{
|
|
Nonce: nonce,
|
|
GasTipCap: common.Big1,
|
|
GasFeeCap: s.chain.Head().BaseFee(),
|
|
To: &common.Address{0xaa},
|
|
Data: randBuf(128),
|
|
Gas: 5_000_000,
|
|
},
|
|
}
|
|
|
|
var txs []*types.Transaction
|
|
for _, inner := range inners {
|
|
tx, err := s.chain.SignTx(from, types.NewTx(inner))
|
|
if err != nil {
|
|
t.Fatalf("failed to sign tx: %v", err)
|
|
}
|
|
txs = append(txs, tx)
|
|
}
|
|
if err := s.sendInvalidTxs(t, txs); err != nil {
|
|
t.Fatalf("failed to send invalid txs: %v", err)
|
|
}
|
|
}
|
|
|
|
func (s *Suite) TestLargeTxRequest(t *utesting.T) {
|
|
t.Log(`This test first send ~2000 transactions to the node, then requests them
|
|
on another peer connection using GetPooledTransactions.`)
|
|
|
|
// Nudge client out of syncing mode to accept pending txs.
|
|
if err := s.engine.sendForkchoiceUpdated(); err != nil {
|
|
t.Fatalf("failed to send next block: %v", err)
|
|
}
|
|
|
|
// Generate many transactions to seed target with.
|
|
var (
|
|
from, nonce = s.chain.GetSender(1)
|
|
count = 2000
|
|
txs []*types.Transaction
|
|
hashes []common.Hash
|
|
set = make(map[common.Hash]struct{})
|
|
)
|
|
for i := 0; i < count; i++ {
|
|
inner := &types.DynamicFeeTx{
|
|
ChainID: s.chain.config.ChainID,
|
|
Nonce: nonce + uint64(i),
|
|
GasTipCap: common.Big1,
|
|
GasFeeCap: s.chain.Head().BaseFee(),
|
|
Gas: 75000,
|
|
}
|
|
tx, err := s.chain.SignTx(from, types.NewTx(inner))
|
|
if err != nil {
|
|
t.Fatalf("failed to sign tx: err")
|
|
}
|
|
txs = append(txs, tx)
|
|
set[tx.Hash()] = struct{}{}
|
|
hashes = append(hashes, tx.Hash())
|
|
}
|
|
s.chain.IncNonce(from, uint64(count))
|
|
|
|
// Send txs.
|
|
if err := s.sendTxs(t, txs); err != nil {
|
|
t.Fatalf("failed to send txs: %v", err)
|
|
}
|
|
|
|
// Set up receive connection to ensure node is peered with the receiving
|
|
// connection before tx request is sent.
|
|
conn, err := s.dial()
|
|
if err != nil {
|
|
t.Fatalf("dial failed: %v", err)
|
|
}
|
|
defer conn.Close()
|
|
if err = conn.peer(s.chain, nil); err != nil {
|
|
t.Fatalf("peering failed: %v", err)
|
|
}
|
|
// Create and send pooled tx request.
|
|
req := ð.GetPooledTransactionsPacket{
|
|
RequestId: 1234,
|
|
GetPooledTransactionsRequest: hashes,
|
|
}
|
|
if err = conn.Write(ethProto, eth.GetPooledTransactionsMsg, req); err != nil {
|
|
t.Fatalf("could not write to conn: %v", err)
|
|
}
|
|
// Check that all received transactions match those that were sent to node.
|
|
msg := new(eth.PooledTransactionsPacket)
|
|
if err := conn.ReadMsg(ethProto, eth.PooledTransactionsMsg, &msg); err != nil {
|
|
t.Fatalf("error reading from connection: %v", err)
|
|
}
|
|
if got, want := msg.RequestId, req.RequestId; got != want {
|
|
t.Fatalf("unexpected request id in response: got %d, want %d", got, want)
|
|
}
|
|
for _, got := range msg.PooledTransactionsResponse {
|
|
if _, exists := set[got.Hash()]; !exists {
|
|
t.Fatalf("unexpected tx received: %v", got.Hash())
|
|
}
|
|
}
|
|
}
|
|
|
|
func (s *Suite) TestNewPooledTxs(t *utesting.T) {
|
|
t.Log(`This test announces transaction hashes to the node and expects it to fetch
|
|
the transactions using a GetPooledTransactions request.`)
|
|
|
|
// Nudge client out of syncing mode to accept pending txs.
|
|
if err := s.engine.sendForkchoiceUpdated(); err != nil {
|
|
t.Fatalf("failed to send next block: %v", err)
|
|
}
|
|
|
|
var (
|
|
count = 50
|
|
from, nonce = s.chain.GetSender(1)
|
|
hashes = make([]common.Hash, count)
|
|
txTypes = make([]byte, count)
|
|
sizes = make([]uint32, count)
|
|
)
|
|
for i := 0; i < count; i++ {
|
|
inner := &types.DynamicFeeTx{
|
|
ChainID: s.chain.config.ChainID,
|
|
Nonce: nonce + uint64(i),
|
|
GasTipCap: common.Big1,
|
|
GasFeeCap: s.chain.Head().BaseFee(),
|
|
Gas: 75000,
|
|
}
|
|
tx, err := s.chain.SignTx(from, types.NewTx(inner))
|
|
if err != nil {
|
|
t.Fatalf("failed to sign tx: err")
|
|
}
|
|
hashes[i] = tx.Hash()
|
|
txTypes[i] = tx.Type()
|
|
sizes[i] = uint32(tx.Size())
|
|
}
|
|
s.chain.IncNonce(from, uint64(count))
|
|
|
|
// Connect to peer.
|
|
conn, err := s.dial()
|
|
if err != nil {
|
|
t.Fatalf("dial failed: %v", err)
|
|
}
|
|
defer conn.Close()
|
|
if err = conn.peer(s.chain, nil); err != nil {
|
|
t.Fatalf("peering failed: %v", err)
|
|
}
|
|
|
|
// Send announcement.
|
|
ann := eth.NewPooledTransactionHashesPacket{Types: txTypes, Sizes: sizes, Hashes: hashes}
|
|
err = conn.Write(ethProto, eth.NewPooledTransactionHashesMsg, ann)
|
|
if err != nil {
|
|
t.Fatalf("failed to write to connection: %v", err)
|
|
}
|
|
|
|
// Wait for GetPooledTxs request.
|
|
for {
|
|
msg, err := conn.ReadEth()
|
|
if err != nil {
|
|
t.Fatalf("failed to read eth msg: %v", err)
|
|
}
|
|
switch msg := msg.(type) {
|
|
case *eth.GetPooledTransactionsPacket:
|
|
if len(msg.GetPooledTransactionsRequest) != len(hashes) {
|
|
t.Fatalf("unexpected number of txs requested: wanted %d, got %d", len(hashes), len(msg.GetPooledTransactionsRequest))
|
|
}
|
|
return
|
|
case *eth.NewPooledTransactionHashesPacket:
|
|
continue
|
|
case *eth.TransactionsPacket:
|
|
continue
|
|
default:
|
|
t.Fatalf("unexpected %s", pretty.Sdump(msg))
|
|
}
|
|
}
|
|
}
|
|
|
|
func makeSidecar(data ...byte) *types.BlobTxSidecar {
|
|
var (
|
|
blobs = make([]kzg4844.Blob, len(data))
|
|
commitments []kzg4844.Commitment
|
|
proofs []kzg4844.Proof
|
|
)
|
|
for i := range blobs {
|
|
blobs[i][0] = data[i]
|
|
c, _ := kzg4844.BlobToCommitment(&blobs[i])
|
|
p, _ := kzg4844.ComputeBlobProof(&blobs[i], c)
|
|
commitments = append(commitments, c)
|
|
proofs = append(proofs, p)
|
|
}
|
|
return &types.BlobTxSidecar{
|
|
Blobs: blobs,
|
|
Commitments: commitments,
|
|
Proofs: proofs,
|
|
}
|
|
}
|
|
|
|
func (s *Suite) makeBlobTxs(count, blobs int, discriminator byte) (txs types.Transactions) {
|
|
from, nonce := s.chain.GetSender(5)
|
|
for i := 0; i < count; i++ {
|
|
// Make blob data, max of 2 blobs per tx.
|
|
blobdata := make([]byte, blobs%3)
|
|
for i := range blobdata {
|
|
blobdata[i] = discriminator
|
|
blobs -= 1
|
|
}
|
|
inner := &types.BlobTx{
|
|
ChainID: uint256.MustFromBig(s.chain.config.ChainID),
|
|
Nonce: nonce + uint64(i),
|
|
GasTipCap: uint256.NewInt(1),
|
|
GasFeeCap: uint256.MustFromBig(s.chain.Head().BaseFee()),
|
|
Gas: 100000,
|
|
BlobFeeCap: uint256.MustFromBig(eip4844.CalcBlobFee(*s.chain.Head().ExcessBlobGas())),
|
|
BlobHashes: makeSidecar(blobdata...).BlobHashes(),
|
|
Sidecar: makeSidecar(blobdata...),
|
|
}
|
|
tx, err := s.chain.SignTx(from, types.NewTx(inner))
|
|
if err != nil {
|
|
panic("blob tx signing failed")
|
|
}
|
|
txs = append(txs, tx)
|
|
}
|
|
return txs
|
|
}
|
|
|
|
func (s *Suite) TestBlobViolations(t *utesting.T) {
|
|
t.Log(`This test sends some invalid blob tx announcements and expects the node to disconnect.`)
|
|
|
|
if err := s.engine.sendForkchoiceUpdated(); err != nil {
|
|
t.Fatalf("send fcu failed: %v", err)
|
|
}
|
|
// Create blob txs for each tests with unique tx hashes.
|
|
var (
|
|
t1 = s.makeBlobTxs(2, 3, 0x1)
|
|
t2 = s.makeBlobTxs(2, 3, 0x2)
|
|
)
|
|
for _, test := range []struct {
|
|
ann eth.NewPooledTransactionHashesPacket
|
|
resp eth.PooledTransactionsResponse
|
|
}{
|
|
// Invalid tx size.
|
|
{
|
|
ann: eth.NewPooledTransactionHashesPacket{
|
|
Types: []byte{types.BlobTxType, types.BlobTxType},
|
|
Sizes: []uint32{uint32(t1[0].Size()), uint32(t1[1].Size() + 10)},
|
|
Hashes: []common.Hash{t1[0].Hash(), t1[1].Hash()},
|
|
},
|
|
resp: eth.PooledTransactionsResponse(t1),
|
|
},
|
|
// Wrong tx type.
|
|
{
|
|
ann: eth.NewPooledTransactionHashesPacket{
|
|
Types: []byte{types.DynamicFeeTxType, types.BlobTxType},
|
|
Sizes: []uint32{uint32(t2[0].Size()), uint32(t2[1].Size())},
|
|
Hashes: []common.Hash{t2[0].Hash(), t2[1].Hash()},
|
|
},
|
|
resp: eth.PooledTransactionsResponse(t2),
|
|
},
|
|
} {
|
|
conn, err := s.dial()
|
|
if err != nil {
|
|
t.Fatalf("dial fail: %v", err)
|
|
}
|
|
if err := conn.peer(s.chain, nil); err != nil {
|
|
t.Fatalf("peering failed: %v", err)
|
|
}
|
|
if err := conn.Write(ethProto, eth.NewPooledTransactionHashesMsg, test.ann); err != nil {
|
|
t.Fatalf("sending announcement failed: %v", err)
|
|
}
|
|
req := new(eth.GetPooledTransactionsPacket)
|
|
if err := conn.ReadMsg(ethProto, eth.GetPooledTransactionsMsg, req); err != nil {
|
|
t.Fatalf("reading pooled tx request failed: %v", err)
|
|
}
|
|
resp := eth.PooledTransactionsPacket{RequestId: req.RequestId, PooledTransactionsResponse: test.resp}
|
|
if err := conn.Write(ethProto, eth.PooledTransactionsMsg, resp); err != nil {
|
|
t.Fatalf("writing pooled tx response failed: %v", err)
|
|
}
|
|
if code, _, err := conn.Read(); err != nil {
|
|
t.Fatalf("expected disconnect on blob violation, got err: %v", err)
|
|
} else if code != discMsg {
|
|
if code == protoOffset(ethProto)+eth.NewPooledTransactionHashesMsg {
|
|
// sometimes we'll get a blob transaction hashes announcement before the disconnect
|
|
// because blob transactions are scheduled to be fetched right away.
|
|
if code, _, err = conn.Read(); err != nil {
|
|
t.Fatalf("expected disconnect on blob violation, got err on second read: %v", err)
|
|
}
|
|
}
|
|
if code != discMsg {
|
|
t.Fatalf("expected disconnect on blob violation, got msg code: %d", code)
|
|
}
|
|
}
|
|
conn.Close()
|
|
}
|
|
}
|