mirror of
https://github.com/ethereum/go-ethereum.git
synced 2026-05-14 20:16:36 +00:00
all: fix problematic function name in comment (#32513)
Fix problematic function name in comment. Do my best to correct them all with a script to avoid spamming PRs.
This commit is contained in:
parent
0979c6a1fa
commit
2a795c14f4
11 changed files with 12 additions and 12 deletions
|
|
@ -269,7 +269,7 @@ func (s *Scheduler) addEvent(event Event) {
|
||||||
s.Trigger()
|
s.Trigger()
|
||||||
}
|
}
|
||||||
|
|
||||||
// filterEvent sorts each Event either as a request event or a server event,
|
// filterEvents sorts each Event either as a request event or a server event,
|
||||||
// depending on its type. Request events are also sorted in a map based on the
|
// depending on its type. Request events are also sorted in a map based on the
|
||||||
// module that originally initiated the request. It also ensures that no events
|
// module that originally initiated the request. It also ensures that no events
|
||||||
// related to a server are returned before EvRegistered or after EvUnregistered.
|
// related to a server are returned before EvRegistered or after EvUnregistered.
|
||||||
|
|
|
||||||
|
|
@ -30,7 +30,7 @@ const (
|
||||||
headLogDelay = time.Second // head indexing log info delay (do not log if finished faster)
|
headLogDelay = time.Second // head indexing log info delay (do not log if finished faster)
|
||||||
)
|
)
|
||||||
|
|
||||||
// updateLoop initializes and updates the log index structure according to the
|
// indexerLoop initializes and updates the log index structure according to the
|
||||||
// current targetView.
|
// current targetView.
|
||||||
func (f *FilterMaps) indexerLoop() {
|
func (f *FilterMaps) indexerLoop() {
|
||||||
defer f.closeWg.Done()
|
defer f.closeWg.Done()
|
||||||
|
|
@ -221,7 +221,7 @@ func (f *FilterMaps) processSingleEvent(blocking bool) bool {
|
||||||
return true
|
return true
|
||||||
}
|
}
|
||||||
|
|
||||||
// setTargetView updates the target chain view of the iterator.
|
// setTarget updates the target chain view of the iterator.
|
||||||
func (f *FilterMaps) setTarget(target targetUpdate) {
|
func (f *FilterMaps) setTarget(target targetUpdate) {
|
||||||
f.targetView = target.targetView
|
f.targetView = target.targetView
|
||||||
f.historyCutoff = target.historyCutoff
|
f.historyCutoff = target.historyCutoff
|
||||||
|
|
|
||||||
|
|
@ -232,7 +232,7 @@ type trieReader struct {
|
||||||
lock sync.Mutex // Lock for protecting concurrent read
|
lock sync.Mutex // Lock for protecting concurrent read
|
||||||
}
|
}
|
||||||
|
|
||||||
// trieReader constructs a trie reader of the specific state. An error will be
|
// newTrieReader constructs a trie reader of the specific state. An error will be
|
||||||
// returned if the associated trie specified by root is not existent.
|
// returned if the associated trie specified by root is not existent.
|
||||||
func newTrieReader(root common.Hash, db *triedb.Database, cache *utils.PointCache) (*trieReader, error) {
|
func newTrieReader(root common.Hash, db *triedb.Database, cache *utils.PointCache) (*trieReader, error) {
|
||||||
var (
|
var (
|
||||||
|
|
|
||||||
|
|
@ -787,7 +787,7 @@ func TestProcessVerkleSelfDestructInSeparateTx(t *testing.T) {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// TestProcessVerkleSelfDestructInSeparateTx controls the contents of the witness after
|
// TestProcessVerkleSelfDestructInSameTx controls the contents of the witness after
|
||||||
// a eip6780-compliant selfdestruct occurs.
|
// a eip6780-compliant selfdestruct occurs.
|
||||||
func TestProcessVerkleSelfDestructInSameTx(t *testing.T) {
|
func TestProcessVerkleSelfDestructInSameTx(t *testing.T) {
|
||||||
// The test txs were taken from a secondary testnet with chain id 69421
|
// The test txs were taken from a secondary testnet with chain id 69421
|
||||||
|
|
|
||||||
|
|
@ -53,7 +53,7 @@ func (p *Program) add(op byte) *Program {
|
||||||
return p
|
return p
|
||||||
}
|
}
|
||||||
|
|
||||||
// pushBig creates a PUSHX instruction and pushes the given val.
|
// doPush creates a PUSHX instruction and pushes the given val.
|
||||||
// - If the val is nil, it pushes zero
|
// - If the val is nil, it pushes zero
|
||||||
// - If the val is bigger than 32 bytes, it panics
|
// - If the val is bigger than 32 bytes, it panics
|
||||||
func (p *Program) doPush(val *uint256.Int) {
|
func (p *Program) doPush(val *uint256.Int) {
|
||||||
|
|
|
||||||
|
|
@ -150,7 +150,7 @@ func ckzgComputeCellProofs(blob *Blob) ([]Proof, error) {
|
||||||
return p, nil
|
return p, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// ckzgVerifyCellProofs verifies that the blob data corresponds to the provided commitment.
|
// ckzgVerifyCellProofBatch verifies that the blob data corresponds to the provided commitment.
|
||||||
func ckzgVerifyCellProofBatch(blobs []Blob, commitments []Commitment, cellProofs []Proof) error {
|
func ckzgVerifyCellProofBatch(blobs []Blob, commitments []Commitment, cellProofs []Proof) error {
|
||||||
ckzgIniter.Do(ckzgInit)
|
ckzgIniter.Do(ckzgInit)
|
||||||
var (
|
var (
|
||||||
|
|
|
||||||
|
|
@ -115,7 +115,7 @@ func gokzgComputeCellProofs(blob *Blob) ([]Proof, error) {
|
||||||
return p, nil
|
return p, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// gokzgVerifyCellProofs verifies that the blob data corresponds to the provided commitment.
|
// gokzgVerifyCellProofBatch verifies that the blob data corresponds to the provided commitment.
|
||||||
func gokzgVerifyCellProofBatch(blobs []Blob, commitments []Commitment, cellProofs []Proof) error {
|
func gokzgVerifyCellProofBatch(blobs []Blob, commitments []Commitment, cellProofs []Proof) error {
|
||||||
gokzgIniter.Do(gokzgInit)
|
gokzgIniter.Do(gokzgInit)
|
||||||
|
|
||||||
|
|
|
||||||
|
|
@ -579,7 +579,7 @@ func newBlockRangeState(chain *core.BlockChain, typeMux *event.TypeMux) *blockRa
|
||||||
return st
|
return st
|
||||||
}
|
}
|
||||||
|
|
||||||
// blockRangeBroadcastLoop announces changes in locally-available block range to peers.
|
// blockRangeLoop announces changes in locally-available block range to peers.
|
||||||
// The range to announce is the range that is available in the store, so it's not just
|
// The range to announce is the range that is available in the store, so it's not just
|
||||||
// about imported blocks.
|
// about imported blocks.
|
||||||
func (h *handler) blockRangeLoop(st *blockRangeState) {
|
func (h *handler) blockRangeLoop(st *blockRangeState) {
|
||||||
|
|
|
||||||
|
|
@ -38,7 +38,7 @@ type SourceIterator interface {
|
||||||
NodeSource() string // source of current node
|
NodeSource() string // source of current node
|
||||||
}
|
}
|
||||||
|
|
||||||
// WithSource attaches a 'source name' to an iterator.
|
// WithSourceName attaches a 'source name' to an iterator.
|
||||||
func WithSourceName(name string, it Iterator) SourceIterator {
|
func WithSourceName(name string, it Iterator) SourceIterator {
|
||||||
return sourceIter{it, name}
|
return sourceIter{it, name}
|
||||||
}
|
}
|
||||||
|
|
|
||||||
|
|
@ -309,7 +309,7 @@ type diskStorageIterator struct {
|
||||||
it ethdb.Iterator
|
it ethdb.Iterator
|
||||||
}
|
}
|
||||||
|
|
||||||
// StorageIterator creates a storage iterator over the persistent state.
|
// newDiskStorageIterator creates a storage iterator over the persistent state.
|
||||||
func newDiskStorageIterator(db ethdb.KeyValueStore, account common.Hash, seek common.Hash) StorageIterator {
|
func newDiskStorageIterator(db ethdb.KeyValueStore, account common.Hash, seek common.Hash) StorageIterator {
|
||||||
pos := common.TrimRightZeroes(seek[:])
|
pos := common.TrimRightZeroes(seek[:])
|
||||||
return &diskStorageIterator{
|
return &diskStorageIterator{
|
||||||
|
|
|
||||||
|
|
@ -181,7 +181,7 @@ func (s *stateSet) accountList() []common.Hash {
|
||||||
return list
|
return list
|
||||||
}
|
}
|
||||||
|
|
||||||
// StorageList returns a sorted list of all storage slot hashes in this state set
|
// storageList returns a sorted list of all storage slot hashes in this state set
|
||||||
// for the given account. The returned list will include the hash of deleted
|
// for the given account. The returned list will include the hash of deleted
|
||||||
// storage slot.
|
// storage slot.
|
||||||
//
|
//
|
||||||
|
|
|
||||||
Loading…
Reference in a new issue