// Copyright 2018 The go-ethereum Authors // This file is part of the go-ethereum library. // // The go-ethereum library is free software: you can redistribute it and/or modify // it under the terms of the GNU Lesser General Public License as published by // the Free Software Foundation, either version 3 of the License, or // (at your option) any later version. // // The go-ethereum library is distributed in the hope that it will be useful, // but WITHOUT ANY WARRANTY; without even the implied warranty of // MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the // GNU Lesser General Public License for more details. // // You should have received a copy of the GNU Lesser General Public License // along with the go-ethereum library. If not, see . package flowcontrol import ( "math/rand" "testing" "time" "github.com/ethereum/go-ethereum/common/mclock" ) type testNode struct { node *ClientNode bufLimit, capacity uint64 waitUntil mclock.AbsTime index, totalCost uint64 } const ( testMaxCost = 1000000 testLength = 100000 ) // testConstantTotalCapacity simulates multiple request sender nodes and verifies // whether the total amount of served requests matches the expected value based on // the total capacity and the duration of the test. // Some nodes are sending requests occasionally so that their buffer should regularly // reach the maximum while other nodes (the "max capacity nodes") are sending at the // maximum permitted rate. The max capacity nodes are changed multiple times during // a single test. func TestConstantTotalCapacity(t *testing.T) { testConstantTotalCapacity(t, 10, 1, 0) testConstantTotalCapacity(t, 10, 1, 1) testConstantTotalCapacity(t, 30, 1, 0) testConstantTotalCapacity(t, 30, 2, 3) testConstantTotalCapacity(t, 100, 1, 0) testConstantTotalCapacity(t, 100, 3, 5) testConstantTotalCapacity(t, 100, 5, 10) } func testConstantTotalCapacity(t *testing.T, nodeCount, maxCapacityNodes, randomSend int) { clock := &mclock.Simulated{} nodes := make([]*testNode, nodeCount) var totalCapacity uint64 for i := range nodes { nodes[i] = &testNode{capacity: uint64(50000 + rand.Intn(100000))} totalCapacity += nodes[i].capacity } m := NewClientManager(PieceWiseLinear{{0, totalCapacity}}, clock) for _, n := range nodes { n.bufLimit = n.capacity * 6000 //uint64(2000+rand.Intn(10000)) n.node = NewClientNode(m, ServerParams{BufLimit: n.bufLimit, MinRecharge: n.capacity}) } maxNodes := make([]int, maxCapacityNodes) for i := range maxNodes { // we don't care if some indexes are selected multiple times // in that case we have fewer max nodes maxNodes[i] = rand.Intn(nodeCount) } for i := 0; i < testLength; i++ { now := clock.Now() for _, idx := range maxNodes { for nodes[idx].send(t, now) { } } if rand.Intn(testLength) < maxCapacityNodes*3 { maxNodes[rand.Intn(maxCapacityNodes)] = rand.Intn(nodeCount) } sendCount := randomSend for sendCount > 0 { if nodes[rand.Intn(nodeCount)].send(t, now) { sendCount-- } } clock.Run(time.Millisecond) } var totalCost uint64 for _, n := range nodes { totalCost += n.totalCost } ratio := float64(totalCost) / float64(totalCapacity) / testLength if ratio < 0.98 || ratio > 1.02 { t.Errorf("totalCost/totalCapacity/testLength ratio incorrect (expected: 1, got: %f)", ratio) } } func (n *testNode) send(t *testing.T, now mclock.AbsTime) bool { if now < n.waitUntil { return false } n.index++ if ok, _, _ := n.node.AcceptRequest(0, n.index, testMaxCost); !ok { t.Fatalf("Rejected request after expected waiting time has passed") } rcost := uint64(rand.Int63n(testMaxCost)) bv := n.node.RequestProcessed(0, n.index, testMaxCost, rcost) if bv < testMaxCost { n.waitUntil = now + mclock.AbsTime((testMaxCost-bv)*1001000/n.capacity) } //n.waitUntil = now + mclock.AbsTime(float64(testMaxCost)*1001000/float64(n.capacity)*(1-float64(bv)/float64(n.bufLimit))) n.totalCost += rcost return true }