serverpool.go 19 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473
  1. // Copyright 2020 The go-ethereum Authors
  2. // This file is part of the go-ethereum library.
  3. //
  4. // The go-ethereum library is free software: you can redistribute it and/or modify
  5. // it under the terms of the GNU Lesser General Public License as published by
  6. // the Free Software Foundation, either version 3 of the License, or
  7. // (at your option) any later version.
  8. //
  9. // The go-ethereum library is distributed in the hope that it will be useful,
  10. // but WITHOUT ANY WARRANTY; without even the implied warranty of
  11. // MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
  12. // GNU Lesser General Public License for more details.
  13. //
  14. // You should have received a copy of the GNU Lesser General Public License
  15. // along with the go-ethereum library. If not, see <http://www.gnu.org/licenses/>.
  16. package les
  17. import (
  18. "errors"
  19. "math/rand"
  20. "reflect"
  21. "sync"
  22. "sync/atomic"
  23. "time"
  24. "github.com/ethereum/go-ethereum/common/mclock"
  25. "github.com/ethereum/go-ethereum/ethdb"
  26. lpc "github.com/ethereum/go-ethereum/les/lespay/client"
  27. "github.com/ethereum/go-ethereum/les/utils"
  28. "github.com/ethereum/go-ethereum/log"
  29. "github.com/ethereum/go-ethereum/p2p/enode"
  30. "github.com/ethereum/go-ethereum/p2p/enr"
  31. "github.com/ethereum/go-ethereum/p2p/nodestate"
  32. "github.com/ethereum/go-ethereum/rlp"
  33. )
  34. const (
  35. minTimeout = time.Millisecond * 500 // minimum request timeout suggested by the server pool
  36. timeoutRefresh = time.Second * 5 // recalculate timeout if older than this
  37. dialCost = 10000 // cost of a TCP dial (used for known node selection weight calculation)
  38. dialWaitStep = 1.5 // exponential multiplier of redial wait time when no value was provided by the server
  39. queryCost = 500 // cost of a UDP pre-negotiation query
  40. queryWaitStep = 1.02 // exponential multiplier of redial wait time when no value was provided by the server
  41. waitThreshold = time.Hour * 2000 // drop node if waiting time is over the threshold
  42. nodeWeightMul = 1000000 // multiplier constant for node weight calculation
  43. nodeWeightThreshold = 100 // minimum weight for keeping a node in the the known (valuable) set
  44. minRedialWait = 10 // minimum redial wait time in seconds
  45. preNegLimit = 5 // maximum number of simultaneous pre-negotiation queries
  46. maxQueryFails = 100 // number of consecutive UDP query failures before we print a warning
  47. )
  48. // serverPool provides a node iterator for dial candidates. The output is a mix of newly discovered
  49. // nodes, a weighted random selection of known (previously valuable) nodes and trusted/paid nodes.
  50. type serverPool struct {
  51. clock mclock.Clock
  52. unixTime func() int64
  53. db ethdb.KeyValueStore
  54. ns *nodestate.NodeStateMachine
  55. vt *lpc.ValueTracker
  56. mixer *enode.FairMix
  57. mixSources []enode.Iterator
  58. dialIterator enode.Iterator
  59. validSchemes enr.IdentityScheme
  60. trustedURLs []string
  61. fillSet *lpc.FillSet
  62. queryFails uint32
  63. timeoutLock sync.RWMutex
  64. timeout time.Duration
  65. timeWeights lpc.ResponseTimeWeights
  66. timeoutRefreshed mclock.AbsTime
  67. }
  68. // nodeHistory keeps track of dial costs which determine node weight together with the
  69. // service value calculated by lpc.ValueTracker.
  70. type nodeHistory struct {
  71. dialCost utils.ExpiredValue
  72. redialWaitStart, redialWaitEnd int64 // unix time (seconds)
  73. }
  74. type nodeHistoryEnc struct {
  75. DialCost utils.ExpiredValue
  76. RedialWaitStart, RedialWaitEnd uint64
  77. }
  78. // queryFunc sends a pre-negotiation query and blocks until a response arrives or timeout occurs.
  79. // It returns 1 if the remote node has confirmed that connection is possible, 0 if not
  80. // possible and -1 if no response arrived (timeout).
  81. type queryFunc func(*enode.Node) int
  82. var (
  83. serverPoolSetup = &nodestate.Setup{Version: 1}
  84. sfHasValue = serverPoolSetup.NewPersistentFlag("hasValue")
  85. sfQueried = serverPoolSetup.NewFlag("queried")
  86. sfCanDial = serverPoolSetup.NewFlag("canDial")
  87. sfDialing = serverPoolSetup.NewFlag("dialed")
  88. sfWaitDialTimeout = serverPoolSetup.NewFlag("dialTimeout")
  89. sfConnected = serverPoolSetup.NewFlag("connected")
  90. sfRedialWait = serverPoolSetup.NewFlag("redialWait")
  91. sfAlwaysConnect = serverPoolSetup.NewFlag("alwaysConnect")
  92. sfDisableSelection = nodestate.MergeFlags(sfQueried, sfCanDial, sfDialing, sfConnected, sfRedialWait)
  93. sfiNodeHistory = serverPoolSetup.NewPersistentField("nodeHistory", reflect.TypeOf(nodeHistory{}),
  94. func(field interface{}) ([]byte, error) {
  95. if n, ok := field.(nodeHistory); ok {
  96. ne := nodeHistoryEnc{
  97. DialCost: n.dialCost,
  98. RedialWaitStart: uint64(n.redialWaitStart),
  99. RedialWaitEnd: uint64(n.redialWaitEnd),
  100. }
  101. enc, err := rlp.EncodeToBytes(&ne)
  102. return enc, err
  103. } else {
  104. return nil, errors.New("invalid field type")
  105. }
  106. },
  107. func(enc []byte) (interface{}, error) {
  108. var ne nodeHistoryEnc
  109. err := rlp.DecodeBytes(enc, &ne)
  110. n := nodeHistory{
  111. dialCost: ne.DialCost,
  112. redialWaitStart: int64(ne.RedialWaitStart),
  113. redialWaitEnd: int64(ne.RedialWaitEnd),
  114. }
  115. return n, err
  116. },
  117. )
  118. sfiNodeWeight = serverPoolSetup.NewField("nodeWeight", reflect.TypeOf(uint64(0)))
  119. sfiConnectedStats = serverPoolSetup.NewField("connectedStats", reflect.TypeOf(lpc.ResponseTimeStats{}))
  120. )
  121. // newServerPool creates a new server pool
  122. func newServerPool(db ethdb.KeyValueStore, dbKey []byte, vt *lpc.ValueTracker, discovery enode.Iterator, mixTimeout time.Duration, query queryFunc, clock mclock.Clock, trustedURLs []string) *serverPool {
  123. s := &serverPool{
  124. db: db,
  125. clock: clock,
  126. unixTime: func() int64 { return time.Now().Unix() },
  127. validSchemes: enode.ValidSchemes,
  128. trustedURLs: trustedURLs,
  129. vt: vt,
  130. ns: nodestate.NewNodeStateMachine(db, []byte(string(dbKey)+"ns:"), clock, serverPoolSetup),
  131. }
  132. s.recalTimeout()
  133. s.mixer = enode.NewFairMix(mixTimeout)
  134. knownSelector := lpc.NewWrsIterator(s.ns, sfHasValue, sfDisableSelection, sfiNodeWeight)
  135. alwaysConnect := lpc.NewQueueIterator(s.ns, sfAlwaysConnect, sfDisableSelection, true, nil)
  136. s.mixSources = append(s.mixSources, knownSelector)
  137. s.mixSources = append(s.mixSources, alwaysConnect)
  138. if discovery != nil {
  139. s.mixSources = append(s.mixSources, discovery)
  140. }
  141. iter := enode.Iterator(s.mixer)
  142. if query != nil {
  143. iter = s.addPreNegFilter(iter, query)
  144. }
  145. s.dialIterator = enode.Filter(iter, func(node *enode.Node) bool {
  146. s.ns.SetState(node, sfDialing, sfCanDial, 0)
  147. s.ns.SetState(node, sfWaitDialTimeout, nodestate.Flags{}, time.Second*10)
  148. return true
  149. })
  150. s.ns.SubscribeState(nodestate.MergeFlags(sfWaitDialTimeout, sfConnected), func(n *enode.Node, oldState, newState nodestate.Flags) {
  151. if oldState.Equals(sfWaitDialTimeout) && newState.IsEmpty() {
  152. // dial timeout, no connection
  153. s.setRedialWait(n, dialCost, dialWaitStep)
  154. s.ns.SetState(n, nodestate.Flags{}, sfDialing, 0)
  155. }
  156. })
  157. s.ns.AddLogMetrics(sfHasValue, sfDisableSelection, "selectable", nil, nil, serverSelectableGauge)
  158. s.ns.AddLogMetrics(sfDialing, nodestate.Flags{}, "dialed", serverDialedMeter, nil, nil)
  159. s.ns.AddLogMetrics(sfConnected, nodestate.Flags{}, "connected", nil, nil, serverConnectedGauge)
  160. return s
  161. }
  162. // addPreNegFilter installs a node filter mechanism that performs a pre-negotiation query.
  163. // Nodes that are filtered out and does not appear on the output iterator are put back
  164. // into redialWait state.
  165. func (s *serverPool) addPreNegFilter(input enode.Iterator, query queryFunc) enode.Iterator {
  166. s.fillSet = lpc.NewFillSet(s.ns, input, sfQueried)
  167. s.ns.SubscribeState(sfQueried, func(n *enode.Node, oldState, newState nodestate.Flags) {
  168. if newState.Equals(sfQueried) {
  169. fails := atomic.LoadUint32(&s.queryFails)
  170. if fails == maxQueryFails {
  171. log.Warn("UDP pre-negotiation query does not seem to work")
  172. }
  173. if fails > maxQueryFails {
  174. fails = maxQueryFails
  175. }
  176. if rand.Intn(maxQueryFails*2) < int(fails) {
  177. // skip pre-negotiation with increasing chance, max 50%
  178. // this ensures that the client can operate even if UDP is not working at all
  179. s.ns.SetState(n, sfCanDial, nodestate.Flags{}, time.Second*10)
  180. // set canDial before resetting queried so that FillSet will not read more
  181. // candidates unnecessarily
  182. s.ns.SetState(n, nodestate.Flags{}, sfQueried, 0)
  183. return
  184. }
  185. go func() {
  186. q := query(n)
  187. if q == -1 {
  188. atomic.AddUint32(&s.queryFails, 1)
  189. } else {
  190. atomic.StoreUint32(&s.queryFails, 0)
  191. }
  192. if q == 1 {
  193. s.ns.SetState(n, sfCanDial, nodestate.Flags{}, time.Second*10)
  194. } else {
  195. s.setRedialWait(n, queryCost, queryWaitStep)
  196. }
  197. s.ns.SetState(n, nodestate.Flags{}, sfQueried, 0)
  198. }()
  199. }
  200. })
  201. return lpc.NewQueueIterator(s.ns, sfCanDial, nodestate.Flags{}, false, func(waiting bool) {
  202. if waiting {
  203. s.fillSet.SetTarget(preNegLimit)
  204. } else {
  205. s.fillSet.SetTarget(0)
  206. }
  207. })
  208. }
  209. // start starts the server pool. Note that NodeStateMachine should be started first.
  210. func (s *serverPool) start() {
  211. s.ns.Start()
  212. for _, iter := range s.mixSources {
  213. // add sources to mixer at startup because the mixer instantly tries to read them
  214. // which should only happen after NodeStateMachine has been started
  215. s.mixer.AddSource(iter)
  216. }
  217. for _, url := range s.trustedURLs {
  218. if node, err := enode.Parse(s.validSchemes, url); err == nil {
  219. s.ns.SetState(node, sfAlwaysConnect, nodestate.Flags{}, 0)
  220. } else {
  221. log.Error("Invalid trusted server URL", "url", url, "error", err)
  222. }
  223. }
  224. unixTime := s.unixTime()
  225. s.ns.ForEach(sfHasValue, nodestate.Flags{}, func(node *enode.Node, state nodestate.Flags) {
  226. s.calculateWeight(node)
  227. if n, ok := s.ns.GetField(node, sfiNodeHistory).(nodeHistory); ok && n.redialWaitEnd > unixTime {
  228. wait := n.redialWaitEnd - unixTime
  229. lastWait := n.redialWaitEnd - n.redialWaitStart
  230. if wait > lastWait {
  231. // if the time until expiration is larger than the last suggested
  232. // waiting time then the system clock was probably adjusted
  233. wait = lastWait
  234. }
  235. s.ns.SetState(node, sfRedialWait, nodestate.Flags{}, time.Duration(wait)*time.Second)
  236. }
  237. })
  238. }
  239. // stop stops the server pool
  240. func (s *serverPool) stop() {
  241. s.dialIterator.Close()
  242. if s.fillSet != nil {
  243. s.fillSet.Close()
  244. }
  245. s.ns.ForEach(sfConnected, nodestate.Flags{}, func(n *enode.Node, state nodestate.Flags) {
  246. // recalculate weight of connected nodes in order to update hasValue flag if necessary
  247. s.calculateWeight(n)
  248. })
  249. s.ns.Stop()
  250. }
  251. // registerPeer implements serverPeerSubscriber
  252. func (s *serverPool) registerPeer(p *serverPeer) {
  253. s.ns.SetState(p.Node(), sfConnected, sfDialing.Or(sfWaitDialTimeout), 0)
  254. nvt := s.vt.Register(p.ID())
  255. s.ns.SetField(p.Node(), sfiConnectedStats, nvt.RtStats())
  256. p.setValueTracker(s.vt, nvt)
  257. p.updateVtParams()
  258. }
  259. // unregisterPeer implements serverPeerSubscriber
  260. func (s *serverPool) unregisterPeer(p *serverPeer) {
  261. s.setRedialWait(p.Node(), dialCost, dialWaitStep)
  262. s.ns.SetState(p.Node(), nodestate.Flags{}, sfConnected, 0)
  263. s.ns.SetField(p.Node(), sfiConnectedStats, nil)
  264. s.vt.Unregister(p.ID())
  265. p.setValueTracker(nil, nil)
  266. }
  267. // recalTimeout calculates the current recommended timeout. This value is used by
  268. // the client as a "soft timeout" value. It also affects the service value calculation
  269. // of individual nodes.
  270. func (s *serverPool) recalTimeout() {
  271. // Use cached result if possible, avoid recalculating too frequently.
  272. s.timeoutLock.RLock()
  273. refreshed := s.timeoutRefreshed
  274. s.timeoutLock.RUnlock()
  275. now := s.clock.Now()
  276. if refreshed != 0 && time.Duration(now-refreshed) < timeoutRefresh {
  277. return
  278. }
  279. // Cached result is stale, recalculate a new one.
  280. rts := s.vt.RtStats()
  281. // Add a fake statistic here. It is an easy way to initialize with some
  282. // conservative values when the database is new. As soon as we have a
  283. // considerable amount of real stats this small value won't matter.
  284. rts.Add(time.Second*2, 10, s.vt.StatsExpFactor())
  285. // Use either 10% failure rate timeout or twice the median response time
  286. // as the recommended timeout.
  287. timeout := minTimeout
  288. if t := rts.Timeout(0.1); t > timeout {
  289. timeout = t
  290. }
  291. if t := rts.Timeout(0.5) * 2; t > timeout {
  292. timeout = t
  293. }
  294. s.timeoutLock.Lock()
  295. if s.timeout != timeout {
  296. s.timeout = timeout
  297. s.timeWeights = lpc.TimeoutWeights(s.timeout)
  298. suggestedTimeoutGauge.Update(int64(s.timeout / time.Millisecond))
  299. totalValueGauge.Update(int64(rts.Value(s.timeWeights, s.vt.StatsExpFactor())))
  300. }
  301. s.timeoutRefreshed = now
  302. s.timeoutLock.Unlock()
  303. }
  304. // getTimeout returns the recommended request timeout.
  305. func (s *serverPool) getTimeout() time.Duration {
  306. s.recalTimeout()
  307. s.timeoutLock.RLock()
  308. defer s.timeoutLock.RUnlock()
  309. return s.timeout
  310. }
  311. // getTimeoutAndWeight returns the recommended request timeout as well as the
  312. // response time weight which is necessary to calculate service value.
  313. func (s *serverPool) getTimeoutAndWeight() (time.Duration, lpc.ResponseTimeWeights) {
  314. s.recalTimeout()
  315. s.timeoutLock.RLock()
  316. defer s.timeoutLock.RUnlock()
  317. return s.timeout, s.timeWeights
  318. }
  319. // addDialCost adds the given amount of dial cost to the node history and returns the current
  320. // amount of total dial cost
  321. func (s *serverPool) addDialCost(n *nodeHistory, amount int64) uint64 {
  322. logOffset := s.vt.StatsExpirer().LogOffset(s.clock.Now())
  323. if amount > 0 {
  324. n.dialCost.Add(amount, logOffset)
  325. }
  326. totalDialCost := n.dialCost.Value(logOffset)
  327. if totalDialCost < dialCost {
  328. totalDialCost = dialCost
  329. }
  330. return totalDialCost
  331. }
  332. // serviceValue returns the service value accumulated in this session and in total
  333. func (s *serverPool) serviceValue(node *enode.Node) (sessionValue, totalValue float64) {
  334. nvt := s.vt.GetNode(node.ID())
  335. if nvt == nil {
  336. return 0, 0
  337. }
  338. currentStats := nvt.RtStats()
  339. _, timeWeights := s.getTimeoutAndWeight()
  340. expFactor := s.vt.StatsExpFactor()
  341. totalValue = currentStats.Value(timeWeights, expFactor)
  342. if connStats, ok := s.ns.GetField(node, sfiConnectedStats).(lpc.ResponseTimeStats); ok {
  343. diff := currentStats
  344. diff.SubStats(&connStats)
  345. sessionValue = diff.Value(timeWeights, expFactor)
  346. sessionValueMeter.Mark(int64(sessionValue))
  347. }
  348. return
  349. }
  350. // updateWeight calculates the node weight and updates the nodeWeight field and the
  351. // hasValue flag. It also saves the node state if necessary.
  352. func (s *serverPool) updateWeight(node *enode.Node, totalValue float64, totalDialCost uint64) {
  353. weight := uint64(totalValue * nodeWeightMul / float64(totalDialCost))
  354. if weight >= nodeWeightThreshold {
  355. s.ns.SetState(node, sfHasValue, nodestate.Flags{}, 0)
  356. s.ns.SetField(node, sfiNodeWeight, weight)
  357. } else {
  358. s.ns.SetState(node, nodestate.Flags{}, sfHasValue, 0)
  359. s.ns.SetField(node, sfiNodeWeight, nil)
  360. }
  361. s.ns.Persist(node) // saved if node history or hasValue changed
  362. }
  363. // setRedialWait calculates and sets the redialWait timeout based on the service value
  364. // and dial cost accumulated during the last session/attempt and in total.
  365. // The waiting time is raised exponentially if no service value has been received in order
  366. // to prevent dialing an unresponsive node frequently for a very long time just because it
  367. // was useful in the past. It can still be occasionally dialed though and once it provides
  368. // a significant amount of service value again its waiting time is quickly reduced or reset
  369. // to the minimum.
  370. // Note: node weight is also recalculated and updated by this function.
  371. func (s *serverPool) setRedialWait(node *enode.Node, addDialCost int64, waitStep float64) {
  372. n, _ := s.ns.GetField(node, sfiNodeHistory).(nodeHistory)
  373. sessionValue, totalValue := s.serviceValue(node)
  374. totalDialCost := s.addDialCost(&n, addDialCost)
  375. // if the current dial session has yielded at least the average value/dial cost ratio
  376. // then the waiting time should be reset to the minimum. If the session value
  377. // is below average but still positive then timeout is limited to the ratio of
  378. // average / current service value multiplied by the minimum timeout. If the attempt
  379. // was unsuccessful then timeout is raised exponentially without limitation.
  380. // Note: dialCost is used in the formula below even if dial was not attempted at all
  381. // because the pre-negotiation query did not return a positive result. In this case
  382. // the ratio has no meaning anyway and waitFactor is always raised, though in smaller
  383. // steps because queries are cheaper and therefore we can allow more failed attempts.
  384. unixTime := s.unixTime()
  385. plannedTimeout := float64(n.redialWaitEnd - n.redialWaitStart) // last planned redialWait timeout
  386. var actualWait float64 // actual waiting time elapsed
  387. if unixTime > n.redialWaitEnd {
  388. // the planned timeout has elapsed
  389. actualWait = plannedTimeout
  390. } else {
  391. // if the node was redialed earlier then we do not raise the planned timeout
  392. // exponentially because that could lead to the timeout rising very high in
  393. // a short amount of time
  394. // Note that in case of an early redial actualWait also includes the dial
  395. // timeout or connection time of the last attempt but it still serves its
  396. // purpose of preventing the timeout rising quicker than linearly as a function
  397. // of total time elapsed without a successful connection.
  398. actualWait = float64(unixTime - n.redialWaitStart)
  399. }
  400. // raise timeout exponentially if the last planned timeout has elapsed
  401. // (use at least the last planned timeout otherwise)
  402. nextTimeout := actualWait * waitStep
  403. if plannedTimeout > nextTimeout {
  404. nextTimeout = plannedTimeout
  405. }
  406. // we reduce the waiting time if the server has provided service value during the
  407. // connection (but never under the minimum)
  408. a := totalValue * dialCost * float64(minRedialWait)
  409. b := float64(totalDialCost) * sessionValue
  410. if a < b*nextTimeout {
  411. nextTimeout = a / b
  412. }
  413. if nextTimeout < minRedialWait {
  414. nextTimeout = minRedialWait
  415. }
  416. wait := time.Duration(float64(time.Second) * nextTimeout)
  417. if wait < waitThreshold {
  418. n.redialWaitStart = unixTime
  419. n.redialWaitEnd = unixTime + int64(nextTimeout)
  420. s.ns.SetField(node, sfiNodeHistory, n)
  421. s.ns.SetState(node, sfRedialWait, nodestate.Flags{}, wait)
  422. s.updateWeight(node, totalValue, totalDialCost)
  423. } else {
  424. // discard known node statistics if waiting time is very long because the node
  425. // hasn't been responsive for a very long time
  426. s.ns.SetField(node, sfiNodeHistory, nil)
  427. s.ns.SetField(node, sfiNodeWeight, nil)
  428. s.ns.SetState(node, nodestate.Flags{}, sfHasValue, 0)
  429. }
  430. }
  431. // calculateWeight calculates and sets the node weight without altering the node history.
  432. // This function should be called during startup and shutdown only, otherwise setRedialWait
  433. // will keep the weights updated as the underlying statistics are adjusted.
  434. func (s *serverPool) calculateWeight(node *enode.Node) {
  435. n, _ := s.ns.GetField(node, sfiNodeHistory).(nodeHistory)
  436. _, totalValue := s.serviceValue(node)
  437. totalDialCost := s.addDialCost(&n, 0)
  438. s.updateWeight(node, totalValue, totalDialCost)
  439. }