mirror of
https://github.com/ethereum/go-ethereum.git
synced 2026-05-17 05:26:36 +00:00
p2p/connmanager: add connection manager to create some churn
Dropping peers randomly with a slow pace to create some artificial churn. Signed-off-by: Csaba Kiraly <csaba.kiraly@gmail.com>
This commit is contained in:
parent
32f36a6749
commit
136d32d2c0
3 changed files with 185 additions and 0 deletions
173
p2p/connmanager.go
Normal file
173
p2p/connmanager.go
Normal file
|
|
@ -0,0 +1,173 @@
|
||||||
|
// Copyright 2015 The go-ethereum Authors
|
||||||
|
// This file is part of the go-ethereum library.
|
||||||
|
//
|
||||||
|
// The go-ethereum library is free software: you can redistribute it and/or modify
|
||||||
|
// it under the terms of the GNU Lesser General Public License as published by
|
||||||
|
// the Free Software Foundation, either version 3 of the License, or
|
||||||
|
// (at your option) any later version.
|
||||||
|
//
|
||||||
|
// The go-ethereum library is distributed in the hope that it will be useful,
|
||||||
|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
||||||
|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
||||||
|
// GNU Lesser General Public License for more details.
|
||||||
|
//
|
||||||
|
// You should have received a copy of the GNU Lesser General Public License
|
||||||
|
// along with the go-ethereum library. If not, see <http://www.gnu.org/licenses/>.
|
||||||
|
|
||||||
|
package p2p
|
||||||
|
|
||||||
|
import (
|
||||||
|
crand "crypto/rand"
|
||||||
|
"encoding/binary"
|
||||||
|
mrand "math/rand"
|
||||||
|
"time"
|
||||||
|
|
||||||
|
"github.com/ethereum/go-ethereum/common"
|
||||||
|
"github.com/ethereum/go-ethereum/common/mclock"
|
||||||
|
"github.com/ethereum/go-ethereum/log"
|
||||||
|
)
|
||||||
|
|
||||||
|
const (
|
||||||
|
// Interval between peer drop events
|
||||||
|
peerDropInterval = 30 * time.Second
|
||||||
|
// How close to max should we initiate the drop timer. O should be fine,
|
||||||
|
// dropping when no more peers can be added. Larger numbers result in more
|
||||||
|
// aggressive drop behavior.
|
||||||
|
peerDropThreshold = 5
|
||||||
|
)
|
||||||
|
|
||||||
|
// connManager monitors the state of the peer pool and makes changes as follows:
|
||||||
|
// - if the peer count is close to the limit, it drops peers randomly every
|
||||||
|
// peerDropInterval to make space for new peers
|
||||||
|
type connManager struct {
|
||||||
|
connmanConfig
|
||||||
|
peersFunc getPeersFunc
|
||||||
|
|
||||||
|
// the peerDrop timer introduces churn if we are close to limit capacity
|
||||||
|
peerDropTimer *mclock.Alarm
|
||||||
|
addPeerCh chan *conn
|
||||||
|
remPeerCh chan *conn
|
||||||
|
}
|
||||||
|
|
||||||
|
// callback type to get the list of connected peers.
|
||||||
|
type getPeersFunc func() []*Peer
|
||||||
|
|
||||||
|
type connmanConfig struct {
|
||||||
|
maxDialPeers int // maximum number of dialed peers
|
||||||
|
log log.Logger
|
||||||
|
clock mclock.Clock
|
||||||
|
rand *mrand.Rand
|
||||||
|
}
|
||||||
|
|
||||||
|
func (cfg connmanConfig) withDefaults() connmanConfig {
|
||||||
|
if cfg.log == nil {
|
||||||
|
cfg.log = log.Root()
|
||||||
|
}
|
||||||
|
if cfg.clock == nil {
|
||||||
|
cfg.clock = mclock.System{}
|
||||||
|
}
|
||||||
|
if cfg.rand == nil {
|
||||||
|
seedb := make([]byte, 8)
|
||||||
|
crand.Read(seedb)
|
||||||
|
seed := int64(binary.BigEndian.Uint64(seedb))
|
||||||
|
cfg.rand = mrand.New(mrand.NewSource(seed))
|
||||||
|
}
|
||||||
|
return cfg
|
||||||
|
}
|
||||||
|
|
||||||
|
func newConnManager(config connmanConfig, peersFunc getPeersFunc) *connManager {
|
||||||
|
cfg := config.withDefaults()
|
||||||
|
cm := &connManager{
|
||||||
|
connmanConfig: cfg,
|
||||||
|
peerDropTimer: mclock.NewAlarm(cfg.clock),
|
||||||
|
peersFunc: peersFunc,
|
||||||
|
addPeerCh: make(chan *conn),
|
||||||
|
remPeerCh: make(chan *conn),
|
||||||
|
}
|
||||||
|
cm.log.Info("New Connection Manager", "maxDialPeers", cm.maxDialPeers, "threshold", peerDropThreshold, "interval", peerDropInterval)
|
||||||
|
go cm.loop()
|
||||||
|
return cm
|
||||||
|
}
|
||||||
|
|
||||||
|
// stop the connection manager.
|
||||||
|
func (cm *connManager) stop() {
|
||||||
|
cm.peerDropTimer.Stop()
|
||||||
|
}
|
||||||
|
|
||||||
|
// peerAdded notifies about peerset change.
|
||||||
|
func (cm *connManager) peerAdded(c *conn) {
|
||||||
|
cm.addPeerCh <- c
|
||||||
|
}
|
||||||
|
|
||||||
|
// peerRemoved notifies about peerset change.
|
||||||
|
func (cm *connManager) peerRemoved(c *conn) {
|
||||||
|
cm.remPeerCh <- c
|
||||||
|
}
|
||||||
|
|
||||||
|
// filter is a helper function to filter the peerset.
|
||||||
|
func filter[T any](s []T, test func(T) bool) (filtered []T) {
|
||||||
|
for _, a := range s {
|
||||||
|
if test(a) {
|
||||||
|
filtered = append(filtered, a)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
// numDialPeers returns the current number of peers dialed (not inbound).
|
||||||
|
func (cm *connManager) numDialPeers() int {
|
||||||
|
selectDialed := func(p *Peer) bool { return !p.rw.is(inboundConn) }
|
||||||
|
dialed := filter(cm.peersFunc(), selectDialed)
|
||||||
|
return len(dialed)
|
||||||
|
}
|
||||||
|
|
||||||
|
func (cm *connManager) numPeers() (int, int, int) {
|
||||||
|
selectDialed := func(p *Peer) bool { return !p.rw.is(inboundConn) }
|
||||||
|
peers := cm.peersFunc()
|
||||||
|
dialed := filter(peers, selectDialed)
|
||||||
|
return len(peers), len(dialed), len(peers) - len(dialed)
|
||||||
|
}
|
||||||
|
|
||||||
|
// dropRandomPeer selects one of the peers randomly and drops it from the peer pool.
|
||||||
|
func (cm *connManager) dropRandomPeer() bool {
|
||||||
|
peers := cm.peersFunc()
|
||||||
|
droppable := peers
|
||||||
|
if len(droppable) > 0 {
|
||||||
|
p := droppable[cm.rand.Intn(len(droppable))]
|
||||||
|
cm.log.Trace("dropping random peer", "id", p.ID(), "duration", common.PrettyDuration(mclock.Now()-p.created), "peercountbefore", len(peers))
|
||||||
|
p.Disconnect(DiscDropped)
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
|
||||||
|
// loop is the main loop of the connection manager.
|
||||||
|
func (cm *connManager) loop() {
|
||||||
|
|
||||||
|
for {
|
||||||
|
|
||||||
|
select {
|
||||||
|
|
||||||
|
case <-cm.addPeerCh:
|
||||||
|
// check and start timer for peer drop
|
||||||
|
// If a drop was already scheduled, Schedule does nothing.
|
||||||
|
numpeers, out, in := cm.numPeers()
|
||||||
|
cm.log.Trace("addPeerCh", "peers", numpeers, "out", out, "in", in, "maxout", cm.maxDialPeers)
|
||||||
|
if cm.maxDialPeers-cm.numDialPeers() <= peerDropThreshold {
|
||||||
|
cm.peerDropTimer.Schedule(cm.clock.Now().Add(peerDropInterval))
|
||||||
|
}
|
||||||
|
|
||||||
|
case <-cm.remPeerCh:
|
||||||
|
// check and stop timer for peer drop
|
||||||
|
numpeers, out, in := cm.numPeers()
|
||||||
|
cm.log.Trace("remPeerCh", "peers", numpeers, "out", out, "in", in, "maxout", cm.maxDialPeers)
|
||||||
|
if cm.maxDialPeers-cm.numDialPeers() > peerDropThreshold {
|
||||||
|
cm.peerDropTimer.Stop()
|
||||||
|
}
|
||||||
|
|
||||||
|
case <-cm.peerDropTimer.C():
|
||||||
|
cm.dropRandomPeer()
|
||||||
|
}
|
||||||
|
}
|
||||||
|
cm.log.Warn("Exiting connmanager loop")
|
||||||
|
}
|
||||||
|
|
@ -69,6 +69,7 @@ const (
|
||||||
DiscUnexpectedIdentity
|
DiscUnexpectedIdentity
|
||||||
DiscSelf
|
DiscSelf
|
||||||
DiscReadTimeout
|
DiscReadTimeout
|
||||||
|
DiscDropped
|
||||||
DiscSubprotocolError = DiscReason(0x10)
|
DiscSubprotocolError = DiscReason(0x10)
|
||||||
|
|
||||||
DiscInvalid = 0xff
|
DiscInvalid = 0xff
|
||||||
|
|
@ -87,6 +88,7 @@ var discReasonToString = [...]string{
|
||||||
DiscUnexpectedIdentity: "unexpected identity",
|
DiscUnexpectedIdentity: "unexpected identity",
|
||||||
DiscSelf: "connected to self",
|
DiscSelf: "connected to self",
|
||||||
DiscReadTimeout: "read timeout",
|
DiscReadTimeout: "read timeout",
|
||||||
|
DiscDropped: "dropped to make space for others",
|
||||||
DiscSubprotocolError: "subprotocol error",
|
DiscSubprotocolError: "subprotocol error",
|
||||||
DiscInvalid: "invalid disconnect reason",
|
DiscInvalid: "invalid disconnect reason",
|
||||||
}
|
}
|
||||||
|
|
|
||||||
|
|
@ -97,6 +97,7 @@ type Server struct {
|
||||||
discv5 *discover.UDPv5
|
discv5 *discover.UDPv5
|
||||||
discmix *enode.FairMix
|
discmix *enode.FairMix
|
||||||
dialsched *dialScheduler
|
dialsched *dialScheduler
|
||||||
|
connman *connManager
|
||||||
|
|
||||||
// This is read by the NAT port mapping loop.
|
// This is read by the NAT port mapping loop.
|
||||||
portMappingRegister chan *portMapping
|
portMappingRegister chan *portMapping
|
||||||
|
|
@ -411,6 +412,8 @@ func (srv *Server) Start() (err error) {
|
||||||
}
|
}
|
||||||
srv.setupDialScheduler()
|
srv.setupDialScheduler()
|
||||||
|
|
||||||
|
srv.setupConnManager()
|
||||||
|
|
||||||
srv.loopWG.Add(1)
|
srv.loopWG.Add(1)
|
||||||
go srv.run()
|
go srv.run()
|
||||||
return nil
|
return nil
|
||||||
|
|
@ -527,6 +530,11 @@ func (srv *Server) setupDialScheduler() {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func (srv *Server) setupConnManager() {
|
||||||
|
config := connmanConfig{maxDialPeers: srv.maxDialedConns()}
|
||||||
|
srv.connman = newConnManager(config, srv.Peers)
|
||||||
|
}
|
||||||
|
|
||||||
func (srv *Server) maxInboundConns() int {
|
func (srv *Server) maxInboundConns() int {
|
||||||
return srv.MaxPeers - srv.maxDialedConns()
|
return srv.MaxPeers - srv.maxDialedConns()
|
||||||
}
|
}
|
||||||
|
|
@ -681,6 +689,7 @@ running:
|
||||||
peers[c.node.ID()] = p
|
peers[c.node.ID()] = p
|
||||||
srv.log.Debug("Adding p2p peer", "peercount", len(peers), "id", p.ID(), "conn", c.flags, "addr", p.RemoteAddr(), "name", p.Name())
|
srv.log.Debug("Adding p2p peer", "peercount", len(peers), "id", p.ID(), "conn", c.flags, "addr", p.RemoteAddr(), "name", p.Name())
|
||||||
srv.dialsched.peerAdded(c)
|
srv.dialsched.peerAdded(c)
|
||||||
|
srv.connman.peerAdded(c)
|
||||||
if p.Inbound() {
|
if p.Inbound() {
|
||||||
inboundCount++
|
inboundCount++
|
||||||
serveSuccessMeter.Mark(1)
|
serveSuccessMeter.Mark(1)
|
||||||
|
|
@ -699,6 +708,7 @@ running:
|
||||||
delete(peers, pd.ID())
|
delete(peers, pd.ID())
|
||||||
srv.log.Debug("Removing p2p peer", "peercount", len(peers), "id", pd.ID(), "duration", d, "req", pd.requested, "err", pd.err)
|
srv.log.Debug("Removing p2p peer", "peercount", len(peers), "id", pd.ID(), "duration", d, "req", pd.requested, "err", pd.err)
|
||||||
srv.dialsched.peerRemoved(pd.rw)
|
srv.dialsched.peerRemoved(pd.rw)
|
||||||
|
srv.connman.peerRemoved(pd.rw)
|
||||||
if pd.Inbound() {
|
if pd.Inbound() {
|
||||||
inboundCount--
|
inboundCount--
|
||||||
activeInboundPeerGauge.Dec(1)
|
activeInboundPeerGauge.Dec(1)
|
||||||
|
|
|
||||||
Loading…
Reference in a new issue