mirror of
https://github.com/ethereum/go-ethereum.git
synced 2026-02-26 15:47:21 +00:00
@lightclient feedbacks
This commit is contained in:
parent
3cd75a3bf9
commit
cb08c7e2bb
16 changed files with 23 additions and 23 deletions
|
|
@ -196,7 +196,7 @@ func TextHash(data []byte) []byte {
|
|||
// This gives context to the signed message and prevents signing of transactions.
|
||||
func TextAndHash(data []byte) ([]byte, string) {
|
||||
msg := fmt.Sprintf("\x19Ethereum Signed Message:\n%d%s", len(data), data)
|
||||
hasher := keccak.NewFastKeccak()
|
||||
hasher := keccak.NewLegacyKeccak256()
|
||||
hasher.Write([]byte(msg))
|
||||
return hasher.Sum(nil), msg
|
||||
}
|
||||
|
|
|
|||
|
|
@ -418,7 +418,7 @@ func MakePreState(db ethdb.Database, accounts types.GenesisAlloc, isBintrie bool
|
|||
}
|
||||
|
||||
func rlpHash(x any) (h common.Hash) {
|
||||
hw := keccak.NewFastKeccak()
|
||||
hw := keccak.NewLegacyKeccak256()
|
||||
rlp.Encode(hw, x)
|
||||
hw.Sum(h[:0])
|
||||
return h
|
||||
|
|
|
|||
|
|
@ -271,7 +271,7 @@ func (a *Address) checksumHex() []byte {
|
|||
buf := a.hex()
|
||||
|
||||
// compute checksum
|
||||
sha := keccak.NewFastKeccak()
|
||||
sha := keccak.NewLegacyKeccak256()
|
||||
sha.Write(buf[2:])
|
||||
hash := sha.Sum(nil)
|
||||
for i := 2; i < len(buf); i++ {
|
||||
|
|
|
|||
|
|
@ -527,7 +527,7 @@ func (ethash *Ethash) FinalizeAndAssemble(chain consensus.ChainHeaderReader, hea
|
|||
|
||||
// SealHash returns the hash of a block prior to it being sealed.
|
||||
func (ethash *Ethash) SealHash(header *types.Header) (hash common.Hash) {
|
||||
hasher := keccak.NewFastKeccak()
|
||||
hasher := keccak.NewLegacyKeccak256()
|
||||
|
||||
enc := []interface{}{
|
||||
header.ParentHash,
|
||||
|
|
|
|||
|
|
@ -69,7 +69,7 @@ func TestBodyStorage(t *testing.T) {
|
|||
|
||||
// Create a test body to move around the database and make sure it's really new
|
||||
body := &types.Body{Uncles: []*types.Header{{Extra: []byte("test header")}}}
|
||||
hasher := keccak.NewFastKeccak()
|
||||
hasher := keccak.NewLegacyKeccak256()
|
||||
rlp.Encode(hasher, body)
|
||||
hash := common.BytesToHash(hasher.Sum(nil))
|
||||
|
||||
|
|
|
|||
|
|
@ -147,7 +147,7 @@ func BenchmarkHashing(b *testing.B) {
|
|||
blockRlp, _ = rlp.EncodeToBytes(block)
|
||||
}
|
||||
var got common.Hash
|
||||
var hasher = keccak.NewFastKeccak()
|
||||
var hasher = keccak.NewLegacyKeccak256()
|
||||
b.Run("iteratorhashing", func(b *testing.B) {
|
||||
for b.Loop() {
|
||||
var hash common.Hash
|
||||
|
|
|
|||
|
|
@ -398,7 +398,7 @@ func GenerateBadBlock(parent *types.Block, engine consensus.Engine, txs types.Tr
|
|||
var receipts []*types.Receipt
|
||||
// The post-state result doesn't need to be correct (this is a bad block), but we do need something there
|
||||
// Preferably something unique. So let's use a combo of blocknum + txhash
|
||||
hasher := keccak.NewFastKeccak()
|
||||
hasher := keccak.NewLegacyKeccak256()
|
||||
hasher.Write(header.Number.Bytes())
|
||||
var cumulativeGas uint64
|
||||
var nBlobs int
|
||||
|
|
|
|||
|
|
@ -36,7 +36,7 @@ func ReturnToPool(h keccak.KeccakState) { hasherPool.Put(h) }
|
|||
|
||||
var hasherPool = sync.Pool{
|
||||
New: func() any {
|
||||
return keccak.NewFastKeccak()
|
||||
return keccak.NewLegacyKeccak256()
|
||||
},
|
||||
}
|
||||
|
||||
|
|
|
|||
|
|
@ -15,6 +15,6 @@ const rate = 136 // sponge rate for Keccak-256: (1600 - 2*256) / 8
|
|||
|
||||
var _ KeccakState = (*Hasher)(nil)
|
||||
|
||||
func NewFastKeccak() *Hasher {
|
||||
func NewLegacyKeccak256() *Hasher {
|
||||
return &Hasher{}
|
||||
}
|
||||
|
|
|
|||
|
|
@ -55,7 +55,7 @@ func TestHashing(t *testing.T) {
|
|||
}
|
||||
var want, got string
|
||||
var old = func() {
|
||||
hasher := keccak.NewFastKeccak()
|
||||
hasher := keccak.NewLegacyKeccak256()
|
||||
for i := 0; i < len(bytecodes); i++ {
|
||||
hasher.Reset()
|
||||
hasher.Write(bytecodes[i])
|
||||
|
|
@ -88,7 +88,7 @@ func BenchmarkHashing(b *testing.B) {
|
|||
bytecodes[i] = buf
|
||||
}
|
||||
var old = func() {
|
||||
hasher := keccak.NewFastKeccak()
|
||||
hasher := keccak.NewLegacyKeccak256()
|
||||
for i := 0; i < len(bytecodes); i++ {
|
||||
hasher.Reset()
|
||||
hasher.Write(bytecodes[i])
|
||||
|
|
|
|||
|
|
@ -39,7 +39,7 @@ type testHasher struct {
|
|||
|
||||
// NewHasher returns a new testHasher instance.
|
||||
func NewHasher() *testHasher {
|
||||
return &testHasher{hasher: keccak.NewFastKeccak()}
|
||||
return &testHasher{hasher: keccak.NewLegacyKeccak256()}
|
||||
}
|
||||
|
||||
// Reset resets the hash state.
|
||||
|
|
|
|||
|
|
@ -262,7 +262,7 @@ const (
|
|||
)
|
||||
|
||||
func subdomain(e entry) string {
|
||||
h := keccak.NewFastKeccak()
|
||||
h := keccak.NewLegacyKeccak256()
|
||||
io.WriteString(h, e.String())
|
||||
return b32format.EncodeToString(h.Sum(nil)[:16])
|
||||
}
|
||||
|
|
@ -272,7 +272,7 @@ func (e *rootEntry) String() string {
|
|||
}
|
||||
|
||||
func (e *rootEntry) sigHash() []byte {
|
||||
h := keccak.NewFastKeccak()
|
||||
h := keccak.NewLegacyKeccak256()
|
||||
fmt.Fprintf(h, rootPrefix+" e=%s l=%s seq=%d", e.eroot, e.lroot, e.seq)
|
||||
return h.Sum(nil)
|
||||
}
|
||||
|
|
|
|||
|
|
@ -49,7 +49,7 @@ func SignV4(r *enr.Record, privkey *ecdsa.PrivateKey) error {
|
|||
cpy.Set(enr.ID("v4"))
|
||||
cpy.Set(Secp256k1(privkey.PublicKey))
|
||||
|
||||
h := keccak.NewFastKeccak()
|
||||
h := keccak.NewLegacyKeccak256()
|
||||
rlp.Encode(h, cpy.AppendElements(nil))
|
||||
sig, err := crypto.Sign(h.Sum(nil), privkey)
|
||||
if err != nil {
|
||||
|
|
@ -70,7 +70,7 @@ func (V4ID) Verify(r *enr.Record, sig []byte) error {
|
|||
return errors.New("invalid public key")
|
||||
}
|
||||
|
||||
h := keccak.NewFastKeccak()
|
||||
h := keccak.NewLegacyKeccak256()
|
||||
rlp.Encode(h, r.AppendElements(nil))
|
||||
if !crypto.VerifySignature(entry, h.Sum(nil), sig) {
|
||||
return enr.ErrInvalidSig
|
||||
|
|
|
|||
|
|
@ -486,10 +486,10 @@ func (h *handshakeState) secrets(auth, authResp []byte) (Secrets, error) {
|
|||
}
|
||||
|
||||
// setup sha3 instances for the MACs
|
||||
mac1 := keccak.NewFastKeccak()
|
||||
mac1 := keccak.NewLegacyKeccak256()
|
||||
mac1.Write(xor(s.MAC, h.respNonce))
|
||||
mac1.Write(auth)
|
||||
mac2 := keccak.NewFastKeccak()
|
||||
mac2 := keccak.NewLegacyKeccak256()
|
||||
mac2.Write(xor(s.MAC, h.initNonce))
|
||||
mac2.Write(authResp)
|
||||
if h.initiator {
|
||||
|
|
|
|||
|
|
@ -496,7 +496,7 @@ func (tx *stTransaction) toMessage(ps stPostState, baseFee *big.Int) (*core.Mess
|
|||
}
|
||||
|
||||
func rlpHash(x interface{}) (h common.Hash) {
|
||||
hw := keccak.NewFastKeccak()
|
||||
hw := keccak.NewLegacyKeccak256()
|
||||
rlp.Encode(hw, x)
|
||||
hw.Sum(h[:0])
|
||||
return h
|
||||
|
|
|
|||
|
|
@ -968,7 +968,7 @@ func TestCommitSequenceStackTrie(t *testing.T) {
|
|||
prng := rand.New(rand.NewSource(int64(count)))
|
||||
// This spongeDb is used to check the sequence of disk-db-writes
|
||||
s := &spongeDb{
|
||||
sponge: keccak.NewFastKeccak(),
|
||||
sponge: keccak.NewLegacyKeccak256(),
|
||||
id: "a",
|
||||
values: make(map[string]string),
|
||||
}
|
||||
|
|
@ -977,7 +977,7 @@ func TestCommitSequenceStackTrie(t *testing.T) {
|
|||
|
||||
// Another sponge is used for the stacktrie commits
|
||||
stackTrieSponge := &spongeDb{
|
||||
sponge: keccak.NewFastKeccak(),
|
||||
sponge: keccak.NewLegacyKeccak256(),
|
||||
id: "b",
|
||||
values: make(map[string]string),
|
||||
}
|
||||
|
|
@ -1040,7 +1040,7 @@ func TestCommitSequenceStackTrie(t *testing.T) {
|
|||
// not fit into 32 bytes, rlp-encoded. However, it's still the correct thing to do.
|
||||
func TestCommitSequenceSmallRoot(t *testing.T) {
|
||||
s := &spongeDb{
|
||||
sponge: keccak.NewFastKeccak(),
|
||||
sponge: keccak.NewLegacyKeccak256(),
|
||||
id: "a",
|
||||
values: make(map[string]string),
|
||||
}
|
||||
|
|
@ -1049,7 +1049,7 @@ func TestCommitSequenceSmallRoot(t *testing.T) {
|
|||
|
||||
// Another sponge is used for the stacktrie commits
|
||||
stackTrieSponge := &spongeDb{
|
||||
sponge: keccak.NewFastKeccak(),
|
||||
sponge: keccak.NewLegacyKeccak256(),
|
||||
id: "b",
|
||||
values: make(map[string]string),
|
||||
}
|
||||
|
|
|
|||
Loading…
Reference in a new issue