aboutsummaryrefslogtreecommitdiffstats
path: root/core
diff options
context:
space:
mode:
authorMission Liao <mission.liao@dexon.org>2018-09-20 09:09:37 +0800
committerGitHub <noreply@github.com>2018-09-20 09:09:37 +0800
commit421d72b2d796195178104a0eb1dedf319ba8664c (patch)
treef32f15c167989905494eca1891c3240b80dac1d6 /core
parent37f117d35c6617e1944d45e001e03813a6a278ed (diff)
downloaddexon-consensus-421d72b2d796195178104a0eb1dedf319ba8664c.tar
dexon-consensus-421d72b2d796195178104a0eb1dedf319ba8664c.tar.gz
dexon-consensus-421d72b2d796195178104a0eb1dedf319ba8664c.tar.bz2
dexon-consensus-421d72b2d796195178104a0eb1dedf319ba8664c.tar.lz
dexon-consensus-421d72b2d796195178104a0eb1dedf319ba8664c.tar.xz
dexon-consensus-421d72b2d796195178104a0eb1dedf319ba8664c.tar.zst
dexon-consensus-421d72b2d796195178104a0eb1dedf319ba8664c.zip
Rename validator* to node* (#120)
Diffstat (limited to 'core')
-rw-r--r--core/agreement-state.go2
-rw-r--r--core/agreement-state_test.go76
-rw-r--r--core/agreement.go48
-rw-r--r--core/agreement_test.go32
-rw-r--r--core/blockdb/level-db_test.go8
-rw-r--r--core/blockdb/memory_test.go4
-rw-r--r--core/compaction-chain.go12
-rw-r--r--core/compaction-chain_test.go12
-rw-r--r--core/consensus-timestamp_test.go8
-rw-r--r--core/consensus.go48
-rw-r--r--core/consensus_test.go110
-rw-r--r--core/crypto.go12
-rw-r--r--core/crypto_test.go20
-rw-r--r--core/dkg-tsig-protocol.go56
-rw-r--r--core/dkg-tsig-protocol_test.go116
-rw-r--r--core/interfaces.go6
-rw-r--r--core/leader-selector_test.go4
-rw-r--r--core/negative-ack.go146
-rw-r--r--core/negative-ack_test.go126
-rw-r--r--core/reliable-broadcast.go46
-rw-r--r--core/reliable-broadcast_test.go82
-rw-r--r--core/test/blocks-generator.go157
-rw-r--r--core/test/blocks-generator_test.go44
-rw-r--r--core/test/fake-transport.go28
-rw-r--r--core/test/governance.go28
-rw-r--r--core/test/interface.go12
-rw-r--r--core/test/revealer.go36
-rw-r--r--core/test/revealer_test.go12
-rw-r--r--core/test/scheduler-event.go8
-rw-r--r--core/test/scheduler.go20
-rw-r--r--core/test/scheduler_test.go60
-rw-r--r--core/test/stopper.go42
-rw-r--r--core/test/stopper_test.go34
-rw-r--r--core/test/tcp-transport.go86
-rw-r--r--core/test/transport_test.go64
-rw-r--r--core/test/utils.go10
-rw-r--r--core/total-ordering_test.go182
-rw-r--r--core/types/block.go2
-rw-r--r--core/types/dkg.go8
-rw-r--r--core/types/node.go (renamed from core/types/validator.go)24
-rw-r--r--core/types/vote.go2
-rw-r--r--core/types/witness.go2
42 files changed, 919 insertions, 916 deletions
diff --git a/core/agreement-state.go b/core/agreement-state.go
index d7225d6..fbee21a 100644
--- a/core/agreement-state.go
+++ b/core/agreement-state.go
@@ -242,7 +242,7 @@ func (s *pass2State) nextState() (agreementState, error) {
s.a.defaultBlock = hash
s.a.period++
oldBlock := s.a.blocks[s.a.ID]
- s.a.blocks = map[types.ValidatorID]*types.Block{
+ s.a.blocks = map[types.NodeID]*types.Block{
s.a.ID: oldBlock,
}
}
diff --git a/core/agreement-state_test.go b/core/agreement-state_test.go
index bc8dc8e..79e4102 100644
--- a/core/agreement-state_test.go
+++ b/core/agreement-state_test.go
@@ -31,8 +31,8 @@ import (
type AgreementStateTestSuite struct {
suite.Suite
- ID types.ValidatorID
- prvKey map[types.ValidatorID]crypto.PrivateKey
+ ID types.NodeID
+ prvKey map[types.NodeID]crypto.PrivateKey
voteChan chan *types.Vote
blockChan chan common.Hash
confirmChan chan common.Hash
@@ -67,13 +67,13 @@ func (s *AgreementStateTestSuite) proposeBlock(
}
func (s *AgreementStateTestSuite) prepareVote(
- vID types.ValidatorID, voteType types.VoteType, blockHash common.Hash,
+ nID types.NodeID, voteType types.VoteType, blockHash common.Hash,
period uint64) (
vote *types.Vote) {
- prvKey, exist := s.prvKey[vID]
+ prvKey, exist := s.prvKey[nID]
s.Require().True(exist)
vote = &types.Vote{
- ProposerID: vID,
+ ProposerID: nID,
Type: voteType,
BlockHash: blockHash,
Period: period,
@@ -87,8 +87,8 @@ func (s *AgreementStateTestSuite) prepareVote(
func (s *AgreementStateTestSuite) SetupTest() {
prvKey, err := eth.NewPrivateKey()
s.Require().Nil(err)
- s.ID = types.NewValidatorID(prvKey.PublicKey())
- s.prvKey = map[types.ValidatorID]crypto.PrivateKey{
+ s.ID = types.NewNodeID(prvKey.PublicKey())
+ s.prvKey = map[types.NodeID]crypto.PrivateKey{
s.ID: prvKey,
}
s.voteChan = make(chan *types.Vote, 100)
@@ -97,24 +97,24 @@ func (s *AgreementStateTestSuite) SetupTest() {
s.block = make(map[common.Hash]*types.Block)
}
-func (s *AgreementStateTestSuite) newAgreement(numValidator int) *agreement {
+func (s *AgreementStateTestSuite) newAgreement(numNode int) *agreement {
leader := newGenesisLeaderSelector("I ❤️ DEXON", eth.SigToPub)
blockProposer := func() *types.Block {
return s.proposeBlock(leader)
}
- validators := make(types.ValidatorIDs, numValidator-1)
- for i := range validators {
+ notarySet := make(types.NodeIDs, numNode-1)
+ for i := range notarySet {
prvKey, err := eth.NewPrivateKey()
s.Require().Nil(err)
- validators[i] = types.NewValidatorID(prvKey.PublicKey())
- s.prvKey[validators[i]] = prvKey
+ notarySet[i] = types.NewNodeID(prvKey.PublicKey())
+ s.prvKey[notarySet[i]] = prvKey
}
- validators = append(validators, s.ID)
+ notarySet = append(notarySet, s.ID)
agreement := newAgreement(
s.ID,
&agreementStateTestReceiver{s},
- validators,
+ notarySet,
leader,
eth.SigToPub,
blockProposer,
@@ -144,8 +144,8 @@ func (s *AgreementStateTestSuite) TestPrepareState() {
_, err = state.nextState()
s.Equal(ErrNoEnoughVoteInPrepareState, err)
- for vID := range a.validators {
- vote := s.prepareVote(vID, types.VotePass, common.Hash{}, 1)
+ for nID := range a.notarySet {
+ vote := s.prepareVote(nID, types.VotePass, common.Hash{}, 1)
s.Require().Nil(a.processVote(vote))
}
@@ -159,11 +159,11 @@ func (s *AgreementStateTestSuite) TestPrepareState() {
block := s.proposeBlock(a.data.leader)
prv, err := eth.NewPrivateKey()
s.Require().Nil(err)
- block.ProposerID = types.NewValidatorID(prv.PublicKey())
+ block.ProposerID = types.NewNodeID(prv.PublicKey())
s.Require().Nil(a.data.leader.prepareBlock(block, prv))
s.Require().Nil(a.processBlock(block))
- for vID := range a.validators {
- vote := s.prepareVote(vID, types.VotePass, block.Hash, 2)
+ for nID := range a.notarySet {
+ vote := s.prepareVote(nID, types.VotePass, block.Hash, 2)
s.Require().Nil(a.processVote(vote))
}
@@ -183,7 +183,7 @@ func (s *AgreementStateTestSuite) TestAckState() {
blocks[i] = s.proposeBlock(a.data.leader)
prv, err := eth.NewPrivateKey()
s.Require().Nil(err)
- blocks[i].ProposerID = types.NewValidatorID(prv.PublicKey())
+ blocks[i].ProposerID = types.NewNodeID(prv.PublicKey())
s.Require().Nil(a.data.leader.prepareBlock(blocks[i], prv))
s.Require().Nil(a.processBlock(blocks[i]))
}
@@ -201,8 +201,8 @@ func (s *AgreementStateTestSuite) TestAckState() {
// For period >= 2, if block v equal to {} has more than 2f+1 pass-vote
// in period 1, propose ack-vote for the block having largest potential.
a.data.period = 2
- for vID := range a.validators {
- vote := s.prepareVote(vID, types.VotePass, common.Hash{}, 1)
+ for nID := range a.notarySet {
+ vote := s.prepareVote(nID, types.VotePass, common.Hash{}, 1)
s.Require().Nil(a.processVote(vote))
}
newState, err = state.nextState()
@@ -217,8 +217,8 @@ func (s *AgreementStateTestSuite) TestAckState() {
// in period 1, propose ack-vote for block v.
hash := blocks[0].Hash
a.data.period = 3
- for vID := range a.validators {
- vote := s.prepareVote(vID, types.VotePass, hash, 2)
+ for nID := range a.notarySet {
+ vote := s.prepareVote(nID, types.VotePass, hash, 2)
s.Require().Nil(a.processVote(vote))
}
newState, err = state.nextState()
@@ -241,8 +241,8 @@ func (s *AgreementStateTestSuite) TestConfirmState() {
a.data.period = 1
block := s.proposeBlock(a.data.leader)
s.Require().Nil(a.processBlock(block))
- for vID := range a.validators {
- vote := s.prepareVote(vID, types.VoteAck, block.Hash, 1)
+ for nID := range a.notarySet {
+ vote := s.prepareVote(nID, types.VoteAck, block.Hash, 1)
s.Require().Nil(a.processVote(vote))
}
s.Require().Nil(state.receiveVote())
@@ -265,8 +265,8 @@ func (s *AgreementStateTestSuite) TestConfirmState() {
// If there are 2f+1 ack-vote for block v equal to {},
// no vote should be proposed.
a.data.period = 3
- for vID := range a.validators {
- vote := s.prepareVote(vID, types.VoteAck, common.Hash{}, 3)
+ for nID := range a.notarySet {
+ vote := s.prepareVote(nID, types.VoteAck, common.Hash{}, 3)
s.Require().Nil(a.processVote(vote))
}
s.Require().Nil(state.receiveVote())
@@ -299,8 +299,8 @@ func (s *AgreementStateTestSuite) TestPass1State() {
// Else if period >= 2 and has 2f+1 pass-vote in period-1 for block {},
// propose pass-vote for block {}.
a.data.period = 2
- for vID := range a.validators {
- vote := s.prepareVote(vID, types.VotePass, common.Hash{}, 1)
+ for nID := range a.notarySet {
+ vote := s.prepareVote(nID, types.VotePass, common.Hash{}, 1)
s.Require().Nil(a.processVote(vote))
}
vote = s.prepareVote(s.ID, types.VoteAck, common.Hash{}, 2)
@@ -318,8 +318,8 @@ func (s *AgreementStateTestSuite) TestPass1State() {
block := s.proposeBlock(a.data.leader)
a.data.defaultBlock = block.Hash
hash = common.NewRandomHash()
- for vID := range a.validators {
- vote := s.prepareVote(vID, types.VotePass, hash, 2)
+ for nID := range a.notarySet {
+ vote := s.prepareVote(nID, types.VotePass, hash, 2)
s.Require().Nil(a.processVote(vote))
}
vote = s.prepareVote(s.ID, types.VoteAck, common.Hash{}, 3)
@@ -371,8 +371,8 @@ func (s *AgreementStateTestSuite) TestPass2State() {
// propose pass-vote for v.
block := s.proposeBlock(a.data.leader)
s.Require().Nil(a.processBlock(block))
- for vID := range a.validators {
- vote := s.prepareVote(vID, types.VoteAck, block.Hash, 1)
+ for nID := range a.notarySet {
+ vote := s.prepareVote(nID, types.VoteAck, block.Hash, 1)
s.Require().Nil(a.processVote(vote))
}
s.Require().Nil(state.receiveVote())
@@ -386,8 +386,8 @@ func (s *AgreementStateTestSuite) TestPass2State() {
a = s.newAgreement(4)
state = newPass2State(a.data)
a.data.period = 2
- for vID := range a.validators {
- vote := s.prepareVote(vID, types.VotePass, common.Hash{}, 1)
+ for nID := range a.notarySet {
+ vote := s.prepareVote(nID, types.VotePass, common.Hash{}, 1)
s.Require().Nil(a.processVote(vote))
}
vote := s.prepareVote(s.ID, types.VoteAck, common.Hash{}, 2)
@@ -412,8 +412,8 @@ func (s *AgreementStateTestSuite) TestPass2State() {
a = s.newAgreement(4)
state = newPass2State(a.data)
a.data.period = 1
- for vID := range a.validators {
- vote := s.prepareVote(vID, types.VotePass, common.Hash{}, 1)
+ for nID := range a.notarySet {
+ vote := s.prepareVote(nID, types.VotePass, common.Hash{}, 1)
s.Require().Nil(a.processVote(vote))
}
s.Require().Nil(state.receiveVote())
diff --git a/core/agreement.go b/core/agreement.go
index 86cb6fc..ffc4ba8 100644
--- a/core/agreement.go
+++ b/core/agreement.go
@@ -30,38 +30,38 @@ import (
// Errors for agreement module.
var (
- ErrNotValidator = fmt.Errorf("not a validaotr")
+ ErrNotInNotarySet = fmt.Errorf("not in notary set")
ErrIncorrectVoteSignature = fmt.Errorf("incorrect vote signature")
)
// ErrFork for fork error in agreement.
type ErrFork struct {
- vID types.ValidatorID
+ nID types.NodeID
old, new common.Hash
}
func (e *ErrFork) Error() string {
return fmt.Sprintf("fork is found for %s, old %s, new %s",
- e.vID.String(), e.old, e.new)
+ e.nID.String(), e.old, e.new)
}
// ErrForkVote for fork vote error in agreement.
type ErrForkVote struct {
- vID types.ValidatorID
+ nID types.NodeID
old, new *types.Vote
}
func (e *ErrForkVote) Error() string {
return fmt.Sprintf("fork vote is found for %s, old %s, new %s",
- e.vID.String(), e.old, e.new)
+ e.nID.String(), e.old, e.new)
}
type blockProposerFn func() *types.Block
-func newVoteListMap() []map[types.ValidatorID]*types.Vote {
- listMap := make([]map[types.ValidatorID]*types.Vote, types.MaxVoteType)
+func newVoteListMap() []map[types.NodeID]*types.Vote {
+ listMap := make([]map[types.NodeID]*types.Vote, types.MaxVoteType)
for idx := range listMap {
- listMap[idx] = make(map[types.ValidatorID]*types.Vote)
+ listMap[idx] = make(map[types.NodeID]*types.Vote)
}
return listMap
}
@@ -87,14 +87,14 @@ type pendingVote struct {
type agreementData struct {
recv agreementReceiver
- ID types.ValidatorID
+ ID types.NodeID
leader *leaderSelector
defaultBlock common.Hash
period uint64
requiredVote int
- votes map[uint64][]map[types.ValidatorID]*types.Vote
+ votes map[uint64][]map[types.NodeID]*types.Vote
votesLock sync.RWMutex
- blocks map[types.ValidatorID]*types.Block
+ blocks map[types.NodeID]*types.Block
blocksLock sync.Mutex
blockProposer blockProposerFn
}
@@ -104,7 +104,7 @@ type agreement struct {
state agreementState
data *agreementData
aID *atomic.Value
- validators map[types.ValidatorID]struct{}
+ notarySet map[types.NodeID]struct{}
sigToPub SigToPubFn
hasOutput bool
lock sync.RWMutex
@@ -115,9 +115,9 @@ type agreement struct {
// newAgreement creates a agreement instance.
func newAgreement(
- ID types.ValidatorID,
+ ID types.NodeID,
recv agreementReceiver,
- validators types.ValidatorIDs,
+ notarySet types.NodeIDs,
leader *leaderSelector,
sigToPub SigToPubFn,
blockProposer blockProposerFn) *agreement {
@@ -132,7 +132,7 @@ func newAgreement(
sigToPub: sigToPub,
candidateBlock: make(map[common.Hash]*types.Block),
}
- agreement.restart(validators, types.Position{})
+ agreement.restart(notarySet, types.Position{})
return agreement
}
@@ -144,7 +144,7 @@ func (a *agreement) terminate() {
}
// restart the agreement
-func (a *agreement) restart(validators types.ValidatorIDs, aID types.Position) {
+func (a *agreement) restart(notarySet types.NodeIDs, aID types.Position) {
func() {
a.lock.Lock()
defer a.lock.Unlock()
@@ -152,18 +152,18 @@ func (a *agreement) restart(validators types.ValidatorIDs, aID types.Position) {
defer a.data.votesLock.Unlock()
a.data.blocksLock.Lock()
defer a.data.blocksLock.Unlock()
- a.data.votes = make(map[uint64][]map[types.ValidatorID]*types.Vote)
+ a.data.votes = make(map[uint64][]map[types.NodeID]*types.Vote)
a.data.votes[1] = newVoteListMap()
a.data.period = 1
- a.data.blocks = make(map[types.ValidatorID]*types.Block)
- a.data.requiredVote = len(validators)/3*2 + 1
+ a.data.blocks = make(map[types.NodeID]*types.Block)
+ a.data.requiredVote = len(notarySet)/3*2 + 1
a.data.leader.restart()
a.data.defaultBlock = common.Hash{}
a.hasOutput = false
a.state = newPrepareState(a.data)
- a.validators = make(map[types.ValidatorID]struct{})
- for _, v := range validators {
- a.validators[v] = struct{}{}
+ a.notarySet = make(map[types.NodeID]struct{})
+ for _, v := range notarySet {
+ a.notarySet[v] = struct{}{}
}
a.candidateBlock = make(map[common.Hash]*types.Block)
a.aID.Store(aID)
@@ -232,10 +232,10 @@ func (a *agreement) sanityCheck(vote *types.Vote) error {
if exist := func() bool {
a.lock.RLock()
defer a.lock.RUnlock()
- _, exist := a.validators[vote.ProposerID]
+ _, exist := a.notarySet[vote.ProposerID]
return exist
}(); !exist {
- return ErrNotValidator
+ return ErrNotInNotarySet
}
ok, err := verifyVoteSignature(vote, a.sigToPub)
if err != nil {
diff --git a/core/agreement_test.go b/core/agreement_test.go
index 4f6ec5b..b344dd9 100644
--- a/core/agreement_test.go
+++ b/core/agreement_test.go
@@ -57,8 +57,8 @@ func (s *AgreementTestSuite) proposeBlock(
type AgreementTestSuite struct {
suite.Suite
- ID types.ValidatorID
- prvKey map[types.ValidatorID]crypto.PrivateKey
+ ID types.NodeID
+ prvKey map[types.NodeID]crypto.PrivateKey
voteChan chan *types.Vote
blockChan chan common.Hash
confirmChan chan common.Hash
@@ -69,8 +69,8 @@ type AgreementTestSuite struct {
func (s *AgreementTestSuite) SetupTest() {
prvKey, err := eth.NewPrivateKey()
s.Require().Nil(err)
- s.ID = types.NewValidatorID(prvKey.PublicKey())
- s.prvKey = map[types.ValidatorID]crypto.PrivateKey{
+ s.ID = types.NewNodeID(prvKey.PublicKey())
+ s.prvKey = map[types.NodeID]crypto.PrivateKey{
s.ID: prvKey,
}
s.voteChan = make(chan *types.Vote, 100)
@@ -79,25 +79,25 @@ func (s *AgreementTestSuite) SetupTest() {
s.block = make(map[common.Hash]*types.Block)
}
-func (s *AgreementTestSuite) newAgreement(numValidator int) *agreement {
+func (s *AgreementTestSuite) newAgreement(numNotarySet int) *agreement {
leader := newGenesisLeaderSelector("🖖👽", eth.SigToPub)
agreementIdx := len(s.agreement)
blockProposer := func() *types.Block {
return s.proposeBlock(agreementIdx)
}
- validators := make(types.ValidatorIDs, numValidator-1)
- for i := range validators {
+ notarySet := make(types.NodeIDs, numNotarySet-1)
+ for i := range notarySet {
prvKey, err := eth.NewPrivateKey()
s.Require().Nil(err)
- validators[i] = types.NewValidatorID(prvKey.PublicKey())
- s.prvKey[validators[i]] = prvKey
+ notarySet[i] = types.NewNodeID(prvKey.PublicKey())
+ s.prvKey[notarySet[i]] = prvKey
}
- validators = append(validators, s.ID)
+ notarySet = append(notarySet, s.ID)
agreement := newAgreement(
s.ID,
&agreementTestReceiver{s},
- validators,
+ notarySet,
leader,
eth.SigToPub,
blockProposer,
@@ -116,7 +116,7 @@ func (s *AgreementTestSuite) prepareVote(vote *types.Vote) {
}
func (s *AgreementTestSuite) copyVote(
- vote *types.Vote, proposer types.ValidatorID) *types.Vote {
+ vote *types.Vote, proposer types.NodeID) *types.Vote {
v := vote.Clone()
v.ProposerID = proposer
s.prepareVote(v)
@@ -138,8 +138,8 @@ func (s *AgreementTestSuite) TestSimpleConfirm() {
s.Require().Len(s.voteChan, 1)
vote := <-s.voteChan
s.Equal(types.VoteAck, vote.Type)
- for vID := range s.prvKey {
- v := s.copyVote(vote, vID)
+ for nID := range s.prvKey {
+ v := s.copyVote(vote, nID)
s.Require().NoError(a.processVote(v))
}
a.nextState()
@@ -147,8 +147,8 @@ func (s *AgreementTestSuite) TestSimpleConfirm() {
s.Require().Len(s.voteChan, 1)
vote = <-s.voteChan
s.Equal(types.VoteConfirm, vote.Type)
- for vID := range s.prvKey {
- v := s.copyVote(vote, vID)
+ for nID := range s.prvKey {
+ v := s.copyVote(vote, nID)
s.Require().NoError(a.processVote(v))
}
// We have enough of Confirm-Votes.
diff --git a/core/blockdb/level-db_test.go b/core/blockdb/level-db_test.go
index 06829f0..9cabc4a 100644
--- a/core/blockdb/level-db_test.go
+++ b/core/blockdb/level-db_test.go
@@ -51,9 +51,9 @@ func (s *LevelDBTestSuite) TestBasicUsage() {
s.Equal(ErrBlockDoesNotExist, err)
// Update on an empty database should not success.
- validator1 := types.ValidatorID{Hash: common.NewRandomHash()}
+ node1 := types.NodeID{Hash: common.NewRandomHash()}
block1 := types.Block{
- ProposerID: validator1,
+ ProposerID: node1,
Hash: hash1,
Position: types.Position{
Height: 1,
@@ -78,7 +78,7 @@ func (s *LevelDBTestSuite) TestBasicUsage() {
err = db.Update(queried)
s.Nil(err)
- // Try to get it back via ValidatorID and height.
+ // Try to get it back via NodeID and height.
queried, err = db.Get(block1.Hash)
s.Nil(err)
@@ -100,7 +100,7 @@ func (s *LevelDBTestSuite) TestSyncIndex() {
blocks := [10]types.Block{}
for i := range blocks {
block := types.Block{
- ProposerID: types.ValidatorID{Hash: common.NewRandomHash()},
+ ProposerID: types.NodeID{Hash: common.NewRandomHash()},
Hash: common.NewRandomHash(),
Position: types.Position{
Height: uint64(i),
diff --git a/core/blockdb/memory_test.go b/core/blockdb/memory_test.go
index 9a3cfa2..a77d355 100644
--- a/core/blockdb/memory_test.go
+++ b/core/blockdb/memory_test.go
@@ -29,12 +29,12 @@ import (
type MemBackedBlockDBTestSuite struct {
suite.Suite
- v0 types.ValidatorID
+ v0 types.NodeID
b00, b01, b02 *types.Block
}
func (s *MemBackedBlockDBTestSuite) SetupSuite() {
- s.v0 = types.ValidatorID{Hash: common.NewRandomHash()}
+ s.v0 = types.NodeID{Hash: common.NewRandomHash()}
genesisHash := common.NewRandomHash()
s.b00 = &types.Block{
diff --git a/core/compaction-chain.go b/core/compaction-chain.go
index c4e0f2f..c72cd7b 100644
--- a/core/compaction-chain.go
+++ b/core/compaction-chain.go
@@ -52,7 +52,7 @@ type compactionChain struct {
prevBlockLock sync.RWMutex
prevBlock *types.Block
witnessAcksLock sync.RWMutex
- latestWitnessAcks map[types.ValidatorID]*types.WitnessAck
+ latestWitnessAcks map[types.NodeID]*types.WitnessAck
sigToPub SigToPubFn
}
@@ -63,7 +63,7 @@ func newCompactionChain(
return &compactionChain{
db: db,
pendingAck: make(map[common.Hash]*pendingAck),
- latestWitnessAcks: make(map[types.ValidatorID]*types.WitnessAck),
+ latestWitnessAcks: make(map[types.NodeID]*types.WitnessAck),
sigToPub: sigToPub,
}
}
@@ -83,7 +83,7 @@ func (cc *compactionChain) sanityCheck(
if err != nil {
return err
}
- if witnessAck.ProposerID != types.NewValidatorID(pubKey) {
+ if witnessAck.ProposerID != types.NewNodeID(pubKey) {
return ErrIncorrectWitnessSignature
}
return nil
@@ -123,7 +123,7 @@ func (cc *compactionChain) prepareWitnessAck(prvKey crypto.PrivateKey) (
return
}
witnessAck = &types.WitnessAck{
- ProposerID: types.NewValidatorID(prvKey.PublicKey()),
+ ProposerID: types.NewNodeID(prvKey.PublicKey()),
WitnessBlockHash: lastBlock.Hash,
Signature: sig,
Hash: hash,
@@ -230,10 +230,10 @@ func (cc *compactionChain) processPendingWitnessAcks() {
cc.pendingAck = pendingAck
}
-func (cc *compactionChain) witnessAcks() map[types.ValidatorID]*types.WitnessAck {
+func (cc *compactionChain) witnessAcks() map[types.NodeID]*types.WitnessAck {
cc.witnessAcksLock.RLock()
defer cc.witnessAcksLock.RUnlock()
- acks := make(map[types.ValidatorID]*types.WitnessAck)
+ acks := make(map[types.NodeID]*types.WitnessAck)
for k, v := range cc.latestWitnessAcks {
acks[k] = v.Clone()
}
diff --git a/core/compaction-chain_test.go b/core/compaction-chain_test.go
index 70a368d..5c08798 100644
--- a/core/compaction-chain_test.go
+++ b/core/compaction-chain_test.go
@@ -117,8 +117,8 @@ func (s *CompactionChainTestSuite) TestProcessWitnessAck() {
s.Require().Nil(err)
prv2, err := eth.NewPrivateKey()
s.Require().Nil(err)
- vID1 := types.NewValidatorID(prv1.PublicKey())
- vID2 := types.NewValidatorID(prv2.PublicKey())
+ nID1 := types.NewNodeID(prv1.PublicKey())
+ nID2 := types.NewNodeID(prv2.PublicKey())
witnessAcks1 := []*types.WitnessAck{}
witnessAcks2 := []*types.WitnessAck{}
for _, block := range blocks {
@@ -150,8 +150,8 @@ func (s *CompactionChainTestSuite) TestProcessWitnessAck() {
s.Nil(cc.processWitnessAck(witnessAcks1[3]))
acks := cc.witnessAcks()
- s.Equal(blocks[3].Hash, acks[vID1].WitnessBlockHash)
- s.Equal(blocks[1].Hash, acks[vID2].WitnessBlockHash)
+ s.Equal(blocks[3].Hash, acks[nID1].WitnessBlockHash)
+ s.Equal(blocks[1].Hash, acks[nID2].WitnessBlockHash)
// Test that witnessAck on less Witness.Height should be ignored.
s.Require().Nil(s.db.Put(*blocks[4]))
@@ -162,8 +162,8 @@ func (s *CompactionChainTestSuite) TestProcessWitnessAck() {
s.Nil(cc.processWitnessAck(witnessAcks2[4]))
acks = cc.witnessAcks()
- s.Equal(blocks[5].Hash, acks[vID1].WitnessBlockHash)
- s.Equal(blocks[5].Hash, acks[vID2].WitnessBlockHash)
+ s.Equal(blocks[5].Hash, acks[nID1].WitnessBlockHash)
+ s.Equal(blocks[5].Hash, acks[nID2].WitnessBlockHash)
}
func TestCompactionChain(t *testing.T) {
diff --git a/core/consensus-timestamp_test.go b/core/consensus-timestamp_test.go
index a5f9fb7..be90f41 100644
--- a/core/consensus-timestamp_test.go
+++ b/core/consensus-timestamp_test.go
@@ -91,7 +91,7 @@ func (s *ConsensusTimestampTest) extractTimestamps(
// should have the same result as processing the whole chain at once.
func (s *ConsensusTimestampTest) TestTimestampPartition() {
blockNums := []int{50, 100, 30}
- validatorNum := 19
+ chainNum := 19
sigma := 100 * time.Millisecond
totalTimestamps := make([]time.Time, 0)
ct := newConsensusTimestamp()
@@ -100,7 +100,7 @@ func (s *ConsensusTimestampTest) TestTimestampPartition() {
totalBlockNum += blockNum
}
totalChain := s.generateBlocksWithTimestamp(
- totalBlockNum, validatorNum, time.Second, sigma)
+ totalBlockNum, chainNum, time.Second, sigma)
for _, blockNum := range blockNums {
var chain []*types.Block
chain, totalChain = totalChain[:blockNum], totalChain[blockNum:]
@@ -118,10 +118,10 @@ func (s *ConsensusTimestampTest) TestTimestampPartition() {
}
func (s *ConsensusTimestampTest) TestTimestampIncrease() {
- validatorNum := 19
+ chainNum := 19
sigma := 100 * time.Millisecond
ct := newConsensusTimestamp()
- chain := s.generateBlocksWithTimestamp(1000, validatorNum, time.Second, sigma)
+ chain := s.generateBlocksWithTimestamp(1000, chainNum, time.Second, sigma)
err := ct.processBlocks(chain)
s.Require().NoError(err)
timestamps := s.extractTimestamps(chain)
diff --git a/core/consensus.go b/core/consensus.go
index 7398628..1af66b3 100644
--- a/core/consensus.go
+++ b/core/consensus.go
@@ -47,8 +47,8 @@ func (e *ErrMissingBlockInfo) Error() string {
// Errors for consensus core.
var (
- ErrProposerNotValidator = fmt.Errorf(
- "proposer is not a validator")
+ ErrProposerNotInNotarySet = fmt.Errorf(
+ "proposer is not in notary set")
ErrIncorrectHash = fmt.Errorf(
"hash of block is incorrect")
ErrIncorrectSignature = fmt.Errorf(
@@ -113,7 +113,7 @@ func (recv *consensusReceiver) ConfirmBlock(hash common.Hash) {
// Consensus implements DEXON Consensus algorithm.
type Consensus struct {
- ID types.ValidatorID
+ ID types.NodeID
app Application
gov Governance
config *types.Config
@@ -143,29 +143,29 @@ func NewConsensus(
sigToPub SigToPubFn) *Consensus {
config := gov.GetConfiguration(0)
- validatorSet := gov.GetNotarySet()
+ notarySet := gov.GetNotarySet()
// Setup acking by information returned from Governace.
rb := newReliableBroadcast()
rb.setChainNum(config.NumChains)
- for vID := range validatorSet {
- rb.addValidator(vID)
+ for nID := range notarySet {
+ rb.addNode(nID)
}
// Setup context.
ctx, ctxCancel := context.WithCancel(context.Background())
// Setup sequencer by information returned from Governace.
- var validators types.ValidatorIDs
- for vID := range validatorSet {
- validators = append(validators, vID)
+ var nodes types.NodeIDs
+ for nID := range notarySet {
+ nodes = append(nodes, nID)
}
to := newTotalOrdering(
uint64(config.K),
- uint64(float32(len(validatorSet)-1)*config.PhiRatio+1),
+ uint64(float32(len(notarySet)-1)*config.PhiRatio+1),
config.NumChains)
con := &Consensus{
- ID: types.NewValidatorID(prv.PublicKey()),
+ ID: types.NewNodeID(prv.PublicKey()),
rbModule: rb,
toModule: to,
ctModule: newConsensusTimestamp(),
@@ -199,7 +199,7 @@ func NewConsensus(
con.baModules[chainID] = newAgreement(
con.ID,
con.receivers[chainID],
- validators,
+ nodes,
newGenesisLeaderSelector(con.config.GenesisCRS, con.sigToPub),
con.sigToPub,
blockProposer,
@@ -230,10 +230,10 @@ func (con *Consensus) Run() {
func (con *Consensus) runBA(chainID uint32, tick <-chan struct{}) {
// TODO(jimmy-dexon): move this function inside agreement.
- validatorSet := con.gov.GetNotarySet()
- validators := make(types.ValidatorIDs, 0, len(validatorSet))
- for vID := range validatorSet {
- validators = append(validators, vID)
+ notarySet := con.gov.GetNotarySet()
+ nodes := make(types.NodeIDs, 0, len(notarySet))
+ for nID := range notarySet {
+ nodes = append(nodes, nID)
}
agreement := con.baModules[chainID]
recv := con.receivers[chainID]
@@ -252,13 +252,13 @@ BALoop:
}
select {
case <-recv.restart:
- // TODO(jimmy-dexon): handling change of validator set.
+ // TODO(jimmy-dexon): handling change of notary set.
aID := types.Position{
ShardID: 0,
ChainID: chainID,
Height: con.rbModule.nextHeight(chainID),
}
- agreement.restart(validators, aID)
+ agreement.restart(nodes, aID)
default:
}
err := agreement.nextState()
@@ -275,8 +275,8 @@ func (con *Consensus) RunLegacy() {
chainID := uint32(0)
hashes := make(common.Hashes, 0, len(con.gov.GetNotarySet()))
- for vID := range con.gov.GetNotarySet() {
- hashes = append(hashes, vID.Hash)
+ for nID := range con.gov.GetNotarySet() {
+ hashes = append(hashes, nID.Hash)
}
sort.Sort(hashes)
for i, hash := range hashes {
@@ -498,7 +498,7 @@ func (con *Consensus) ProcessBlock(block *types.Block) (err error) {
func (con *Consensus) checkPrepareBlock(
b *types.Block, proposeTime time.Time) (err error) {
- if (b.ProposerID == types.ValidatorID{}) {
+ if (b.ProposerID == types.NodeID{}) {
err = &ErrMissingBlockInfo{MissingField: "ProposerID"}
return
}
@@ -556,14 +556,14 @@ func (con *Consensus) PrepareGenesisBlock(b *types.Block,
func (con *Consensus) ProcessWitnessAck(witnessAck *types.WitnessAck) (err error) {
witnessAck = witnessAck.Clone()
if _, exists := con.gov.GetNotarySet()[witnessAck.ProposerID]; !exists {
- err = ErrProposerNotValidator
+ err = ErrProposerNotInNotarySet
return
}
err = con.ccModule.processWitnessAck(witnessAck)
return
}
-// WitnessAcks returns the latest WitnessAck received from all other validators.
-func (con *Consensus) WitnessAcks() map[types.ValidatorID]*types.WitnessAck {
+// WitnessAcks returns the latest WitnessAck received from all other nodes.
+func (con *Consensus) WitnessAcks() map[types.NodeID]*types.WitnessAck {
return con.ccModule.witnessAcks()
}
diff --git a/core/consensus_test.go b/core/consensus_test.go
index 32f9986..c3f33fa 100644
--- a/core/consensus_test.go
+++ b/core/consensus_test.go
@@ -47,7 +47,7 @@ func (n *network) BroadcastWitnessAck(witnessAck *types.WitnessAck) {
// SendDKGPrivateShare sends PrivateShare to a DKG participant.
func (n *network) SendDKGPrivateShare(
- recv types.ValidatorID, prvShare *types.DKGPrivateShare) {
+ recv types.NodeID, prvShare *types.DKGPrivateShare) {
}
// ReceiveChan returns a channel to receive messages from DEXON network.
@@ -60,7 +60,7 @@ type ConsensusTestSuite struct {
}
func (s *ConsensusTestSuite) prepareGenesisBlock(
- proposerID types.ValidatorID,
+ proposerID types.NodeID,
chainID uint32,
con *Consensus) *types.Block {
@@ -76,12 +76,12 @@ func (s *ConsensusTestSuite) prepareGenesisBlock(
}
func (s *ConsensusTestSuite) prepareConsensus(
- gov *test.Governance, vID types.ValidatorID) (*Application, *Consensus) {
+ gov *test.Governance, nID types.NodeID) (*Application, *Consensus) {
app := test.NewApp()
db, err := blockdb.NewMemBackedBlockDB()
s.Require().Nil(err)
- prv, exist := gov.GetPrivateKey(vID)
+ prv, exist := gov.GetPrivateKey(nID)
s.Require().Nil(exist)
con := NewConsensus(app, gov, db, &network{}, prv, eth.SigToPub)
return &con.app, con
@@ -96,29 +96,29 @@ func (s *ConsensusTestSuite) TestSimpleDeliverBlock() {
// | o | | <- the only block which is acked by all other blocks
// |/|\|\| at the same height.
// o o o o <- genesis blocks
- // 0 1 2 3 <- index of validator ID
+ // 0 1 2 3 <- index of node ID
//
// This test case only works for Total Ordering with K=0.
var (
minInterval = 50 * time.Millisecond
gov, err = test.NewGovernance(4, time.Second)
req = s.Require()
- validators []types.ValidatorID
+ nodes []types.NodeID
)
s.Require().Nil(err)
- for vID := range gov.GetNotarySet() {
- validators = append(validators, vID)
+ for nID := range gov.GetNotarySet() {
+ nodes = append(nodes, nID)
}
// Setup core.Consensus and test.App.
- objs := map[types.ValidatorID]*struct {
+ objs := map[types.NodeID]*struct {
app *Application
con *Consensus
}{}
- for _, vID := range validators {
- app, con := s.prepareConsensus(gov, vID)
- objs[vID] = &struct {
+ for _, nID := range nodes {
+ app, con := s.prepareConsensus(gov, nID)
+ objs[nID] = &struct {
app *Application
con *Consensus
}{app, con}
@@ -131,13 +131,13 @@ func (s *ConsensusTestSuite) TestSimpleDeliverBlock() {
}
}
// Genesis blocks
- b00 := s.prepareGenesisBlock(validators[0], 0, objs[validators[0]].con)
+ b00 := s.prepareGenesisBlock(nodes[0], 0, objs[nodes[0]].con)
time.Sleep(minInterval)
- b10 := s.prepareGenesisBlock(validators[1], 1, objs[validators[1]].con)
+ b10 := s.prepareGenesisBlock(nodes[1], 1, objs[nodes[1]].con)
time.Sleep(minInterval)
- b20 := s.prepareGenesisBlock(validators[2], 2, objs[validators[2]].con)
+ b20 := s.prepareGenesisBlock(nodes[2], 2, objs[nodes[2]].con)
time.Sleep(minInterval)
- b30 := s.prepareGenesisBlock(validators[3], 3, objs[validators[3]].con)
+ b30 := s.prepareGenesisBlock(nodes[3], 3, objs[nodes[3]].con)
broadcast(b00)
broadcast(b10)
broadcast(b20)
@@ -145,14 +145,14 @@ func (s *ConsensusTestSuite) TestSimpleDeliverBlock() {
// Setup b11.
time.Sleep(minInterval)
b11 := &types.Block{
- ProposerID: validators[1],
+ ProposerID: nodes[1],
Position: types.Position{
ChainID: 1,
},
}
b11.Hash, err = hashBlock(b11)
s.Require().Nil(err)
- req.Nil(objs[validators[1]].con.PrepareBlock(b11, time.Now().UTC()))
+ req.Nil(objs[nodes[1]].con.PrepareBlock(b11, time.Now().UTC()))
req.Len(b11.Acks, 4)
req.Contains(b11.Acks, b00.Hash)
req.Contains(b11.Acks, b10.Hash)
@@ -162,37 +162,37 @@ func (s *ConsensusTestSuite) TestSimpleDeliverBlock() {
// Setup b01.
time.Sleep(minInterval)
b01 := &types.Block{
- ProposerID: validators[0],
+ ProposerID: nodes[0],
Position: types.Position{
ChainID: 0,
},
Hash: common.NewRandomHash(),
}
- req.Nil(objs[validators[0]].con.PrepareBlock(b01, time.Now().UTC()))
+ req.Nil(objs[nodes[0]].con.PrepareBlock(b01, time.Now().UTC()))
req.Len(b01.Acks, 4)
req.Contains(b01.Acks, b11.Hash)
// Setup b21.
time.Sleep(minInterval)
b21 := &types.Block{
- ProposerID: validators[2],
+ ProposerID: nodes[2],
Position: types.Position{
ChainID: 2,
},
Hash: common.NewRandomHash(),
}
- req.Nil(objs[validators[2]].con.PrepareBlock(b21, time.Now().UTC()))
+ req.Nil(objs[nodes[2]].con.PrepareBlock(b21, time.Now().UTC()))
req.Len(b21.Acks, 4)
req.Contains(b21.Acks, b11.Hash)
// Setup b31.
time.Sleep(minInterval)
b31 := &types.Block{
- ProposerID: validators[3],
+ ProposerID: nodes[3],
Position: types.Position{
ChainID: 3,
},
Hash: common.NewRandomHash(),
}
- req.Nil(objs[validators[3]].con.PrepareBlock(b31, time.Now().UTC()))
+ req.Nil(objs[nodes[3]].con.PrepareBlock(b31, time.Now().UTC()))
req.Len(b31.Acks, 4)
req.Contains(b31.Acks, b11.Hash)
// Broadcast other height=1 blocks.
@@ -203,13 +203,13 @@ func (s *ConsensusTestSuite) TestSimpleDeliverBlock() {
// Setup b02.
time.Sleep(minInterval)
b02 := &types.Block{
- ProposerID: validators[0],
+ ProposerID: nodes[0],
Position: types.Position{
ChainID: 0,
},
Hash: common.NewRandomHash(),
}
- req.Nil(objs[validators[0]].con.PrepareBlock(b02, time.Now().UTC()))
+ req.Nil(objs[nodes[0]].con.PrepareBlock(b02, time.Now().UTC()))
req.Len(b02.Acks, 3)
req.Contains(b02.Acks, b01.Hash)
req.Contains(b02.Acks, b21.Hash)
@@ -217,13 +217,13 @@ func (s *ConsensusTestSuite) TestSimpleDeliverBlock() {
// Setup b12.
time.Sleep(minInterval)
b12 := &types.Block{
- ProposerID: validators[1],
+ ProposerID: nodes[1],
Position: types.Position{
ChainID: 1,
},
Hash: common.NewRandomHash(),
}
- req.Nil(objs[validators[1]].con.PrepareBlock(b12, time.Now().UTC()))
+ req.Nil(objs[nodes[1]].con.PrepareBlock(b12, time.Now().UTC()))
req.Len(b12.Acks, 4)
req.Contains(b12.Acks, b01.Hash)
req.Contains(b12.Acks, b11.Hash)
@@ -232,13 +232,13 @@ func (s *ConsensusTestSuite) TestSimpleDeliverBlock() {
// Setup b22.
time.Sleep(minInterval)
b22 := &types.Block{
- ProposerID: validators[2],
+ ProposerID: nodes[2],
Position: types.Position{
ChainID: 2,
},
Hash: common.NewRandomHash(),
}
- req.Nil(objs[validators[2]].con.PrepareBlock(b22, time.Now().UTC()))
+ req.Nil(objs[nodes[2]].con.PrepareBlock(b22, time.Now().UTC()))
req.Len(b22.Acks, 3)
req.Contains(b22.Acks, b01.Hash)
req.Contains(b22.Acks, b21.Hash)
@@ -246,13 +246,13 @@ func (s *ConsensusTestSuite) TestSimpleDeliverBlock() {
// Setup b32.
time.Sleep(minInterval)
b32 := &types.Block{
- ProposerID: validators[3],
+ ProposerID: nodes[3],
Position: types.Position{
ChainID: 3,
},
Hash: common.NewRandomHash(),
}
- req.Nil(objs[validators[3]].con.PrepareBlock(b32, time.Now().UTC()))
+ req.Nil(objs[nodes[3]].con.PrepareBlock(b32, time.Now().UTC()))
req.Len(b32.Acks, 3)
req.Contains(b32.Acks, b01.Hash)
req.Contains(b32.Acks, b21.Hash)
@@ -324,30 +324,30 @@ func (s *ConsensusTestSuite) TestPrepareBlock() {
// - Make sure Consensus.PrepareBlock would only attempt to
// ack the prepared block.
var (
- gov, err = test.NewGovernance(4, time.Second)
- req = s.Require()
- validators []types.ValidatorID
+ gov, err = test.NewGovernance(4, time.Second)
+ req = s.Require()
+ nodes []types.NodeID
)
s.Require().Nil(err)
- for vID := range gov.GetNotarySet() {
- validators = append(validators, vID)
+ for nID := range gov.GetNotarySet() {
+ nodes = append(nodes, nID)
}
// Setup core.Consensus and test.App.
- objs := map[types.ValidatorID]*struct {
+ objs := map[types.NodeID]*struct {
app *Application
con *Consensus
}{}
- for _, vID := range validators {
- app, con := s.prepareConsensus(gov, vID)
- objs[vID] = &struct {
+ for _, nID := range nodes {
+ app, con := s.prepareConsensus(gov, nID)
+ objs[nID] = &struct {
app *Application
con *Consensus
}{app, con}
}
- b00 := s.prepareGenesisBlock(validators[0], 0, objs[validators[0]].con)
- b10 := s.prepareGenesisBlock(validators[1], 1, objs[validators[1]].con)
- b20 := s.prepareGenesisBlock(validators[2], 2, objs[validators[2]].con)
- b30 := s.prepareGenesisBlock(validators[3], 3, objs[validators[3]].con)
+ b00 := s.prepareGenesisBlock(nodes[0], 0, objs[nodes[0]].con)
+ b10 := s.prepareGenesisBlock(nodes[1], 1, objs[nodes[1]].con)
+ b20 := s.prepareGenesisBlock(nodes[2], 2, objs[nodes[2]].con)
+ b30 := s.prepareGenesisBlock(nodes[3], 3, objs[nodes[3]].con)
for _, obj := range objs {
con := obj.con
req.Nil(con.ProcessBlock(b00))
@@ -356,11 +356,11 @@ func (s *ConsensusTestSuite) TestPrepareBlock() {
req.Nil(con.ProcessBlock(b30))
}
b11 := &types.Block{
- ProposerID: validators[1],
+ ProposerID: nodes[1],
}
// Sleep to make sure 'now' is slower than b10's timestamp.
time.Sleep(100 * time.Millisecond)
- req.Nil(objs[validators[1]].con.PrepareBlock(b11, time.Now().UTC()))
+ req.Nil(objs[nodes[1]].con.PrepareBlock(b11, time.Now().UTC()))
// Make sure we would assign 'now' to the timestamp belongs to
// the proposer.
req.True(
@@ -371,25 +371,25 @@ func (s *ConsensusTestSuite) TestPrepareBlock() {
req.Nil(con.ProcessBlock(b11))
}
b12 := &types.Block{
- ProposerID: validators[1],
+ ProposerID: nodes[1],
}
- req.Nil(objs[validators[1]].con.PrepareBlock(b12, time.Now().UTC()))
+ req.Nil(objs[nodes[1]].con.PrepareBlock(b12, time.Now().UTC()))
req.Len(b12.Acks, 1)
req.Contains(b12.Acks, b11.Hash)
}
func (s *ConsensusTestSuite) TestPrepareGenesisBlock() {
var (
- gov, err = test.NewGovernance(4, time.Second)
- validators []types.ValidatorID
+ gov, err = test.NewGovernance(4, time.Second)
+ nodes []types.NodeID
)
s.Require().Nil(err)
- for vID := range gov.GetNotarySet() {
- validators = append(validators, vID)
+ for nID := range gov.GetNotarySet() {
+ nodes = append(nodes, nID)
}
- _, con := s.prepareConsensus(gov, validators[0])
+ _, con := s.prepareConsensus(gov, nodes[0])
block := &types.Block{
- ProposerID: validators[0],
+ ProposerID: nodes[0],
}
con.PrepareGenesisBlock(block, time.Now().UTC())
s.True(block.IsGenesis())
diff --git a/core/crypto.go b/core/crypto.go
index 707a2b9..e68d7cc 100644
--- a/core/crypto.go
+++ b/core/crypto.go
@@ -103,7 +103,7 @@ func verifyVoteSignature(vote *types.Vote, sigToPub SigToPubFn) (bool, error) {
if err != nil {
return false, err
}
- if vote.ProposerID != types.NewValidatorID(pubKey) {
+ if vote.ProposerID != types.NewNodeID(pubKey) {
return false, nil
}
return true, nil
@@ -121,7 +121,7 @@ func verifyCRSSignature(block *types.Block, crs common.Hash, sigToPub SigToPubFn
if err != nil {
return false, err
}
- if block.ProposerID != types.NewValidatorID(pubKey) {
+ if block.ProposerID != types.NewNodeID(pubKey) {
return false, nil
}
return true, nil
@@ -162,7 +162,7 @@ func verifyDKGPrivateShareSignature(
if err != nil {
return false, err
}
- if prvShare.ProposerID != types.NewValidatorID(pubKey) {
+ if prvShare.ProposerID != types.NewNodeID(pubKey) {
return false, nil
}
return true, nil
@@ -187,7 +187,7 @@ func verifyDKGMasterPublicKeySignature(
if err != nil {
return false, err
}
- if mpk.ProposerID != types.NewValidatorID(pubKey) {
+ if mpk.ProposerID != types.NewNodeID(pubKey) {
return false, nil
}
return true, nil
@@ -213,7 +213,7 @@ func verifyDKGComplaintSignature(
if err != nil {
return false, err
}
- if complaint.ProposerID != types.NewValidatorID(pubKey) {
+ if complaint.ProposerID != types.NewNodeID(pubKey) {
return false, nil
}
return true, nil
@@ -237,7 +237,7 @@ func verifyDKGPartialSignatureSignature(
if err != nil {
return false, err
}
- if psig.ProposerID != types.NewValidatorID(pubKey) {
+ if psig.ProposerID != types.NewNodeID(pubKey) {
return false, nil
}
return true, nil
diff --git a/core/crypto_test.go b/core/crypto_test.go
index fe6bbc1..6c807da 100644
--- a/core/crypto_test.go
+++ b/core/crypto_test.go
@@ -33,7 +33,7 @@ type CryptoTestSuite struct {
suite.Suite
}
-var myVID = types.ValidatorID{Hash: common.NewRandomHash()}
+var myNID = types.NodeID{Hash: common.NewRandomHash()}
func (s *CryptoTestSuite) prepareBlock(prevBlock *types.Block) *types.Block {
acks := common.Hashes{}
@@ -181,9 +181,9 @@ func (s *CryptoTestSuite) TestVoteSignature() {
prv, err := eth.NewPrivateKey()
s.Require().Nil(err)
pub := prv.PublicKey()
- vID := types.NewValidatorID(pub)
+ nID := types.NewNodeID(pub)
vote := &types.Vote{
- ProposerID: vID,
+ ProposerID: nID,
Type: types.VoteAck,
BlockHash: common.NewRandomHash(),
Period: 1,
@@ -200,9 +200,9 @@ func (s *CryptoTestSuite) TestCRSSignature() {
prv, err := eth.NewPrivateKey()
s.Require().Nil(err)
pub := prv.PublicKey()
- vID := types.NewValidatorID(pub)
+ nID := types.NewNodeID(pub)
block := &types.Block{
- ProposerID: vID,
+ ProposerID: nID,
}
block.CRSSignature, err = prv.Sign(hashCRS(block, crs))
s.Require().Nil(err)
@@ -214,9 +214,9 @@ func (s *CryptoTestSuite) TestCRSSignature() {
func (s *CryptoTestSuite) TestDKGSignature() {
prv, err := eth.NewPrivateKey()
s.Require().Nil(err)
- vID := types.NewValidatorID(prv.PublicKey())
+ nID := types.NewNodeID(prv.PublicKey())
prvShare := &types.DKGPrivateShare{
- ProposerID: vID,
+ ProposerID: nID,
Round: 5,
PrivateShare: *dkg.NewPrivateKey(),
}
@@ -229,7 +229,7 @@ func (s *CryptoTestSuite) TestDKGSignature() {
id := dkg.NewID([]byte{13})
_, pkShare := dkg.NewPrivateKeyShares(1)
mpk := &types.DKGMasterPublicKey{
- ProposerID: vID,
+ ProposerID: nID,
Round: 5,
DKGID: id,
PublicKeyShares: *pkShare,
@@ -241,7 +241,7 @@ func (s *CryptoTestSuite) TestDKGSignature() {
s.False(verifyDKGMasterPublicKeySignature(mpk, eth.SigToPub))
complaint := &types.DKGComplaint{
- ProposerID: vID,
+ ProposerID: nID,
Round: 5,
PrivateShare: *prvShare,
}
@@ -252,7 +252,7 @@ func (s *CryptoTestSuite) TestDKGSignature() {
s.False(verifyDKGComplaintSignature(complaint, eth.SigToPub))
sig := &types.DKGPartialSignature{
- ProposerID: vID,
+ ProposerID: nID,
Round: 5,
PartialSignature: dkg.PartialSignature{},
}
diff --git a/core/dkg-tsig-protocol.go b/core/dkg-tsig-protocol.go
index 2db18f1..e1a0635 100644
--- a/core/dkg-tsig-protocol.go
+++ b/core/dkg-tsig-protocol.go
@@ -54,20 +54,20 @@ type dkgComplaintReceiver interface {
ProposeDKGMasterPublicKey(mpk *types.DKGMasterPublicKey)
// ProposeDKGPrivateShare propose a DKGPrivateShare.
- ProposeDKGPrivateShare(to types.ValidatorID, prv *types.DKGPrivateShare)
+ ProposeDKGPrivateShare(to types.NodeID, prv *types.DKGPrivateShare)
}
type dkgProtocol struct {
- ID types.ValidatorID
+ ID types.NodeID
recv dkgComplaintReceiver
round uint64
threshold int
sigToPub SigToPubFn
- idMap map[types.ValidatorID]dkg.ID
- mpkMap map[types.ValidatorID]*dkg.PublicKeyShares
+ idMap map[types.NodeID]dkg.ID
+ mpkMap map[types.NodeID]*dkg.PublicKeyShares
masterPrivateShare *dkg.PrivateKeyShares
prvShares *dkg.PrivateKeyShares
- prvSharesReceived map[types.ValidatorID]struct{}
+ prvSharesReceived map[types.NodeID]struct{}
}
type dkgShareSecret struct {
@@ -77,8 +77,8 @@ type dkgShareSecret struct {
type dkgGroupPublicKey struct {
round uint64
qualifyIDs dkg.IDs
- idMap map[types.ValidatorID]dkg.ID
- publicKeys map[types.ValidatorID]*dkg.PublicKey
+ idMap map[types.NodeID]dkg.ID
+ publicKeys map[types.NodeID]*dkg.PublicKey
groupPublicKey *dkg.PublicKey
threshold int
sigToPub SigToPubFn
@@ -90,12 +90,12 @@ type tsigProtocol struct {
threshold int
}
-func newDKGID(ID types.ValidatorID) dkg.ID {
+func newDKGID(ID types.NodeID) dkg.ID {
return dkg.NewID(ID.Hash[:])
}
func newDKGProtocol(
- ID types.ValidatorID,
+ ID types.NodeID,
recv dkgComplaintReceiver,
round uint64,
threshold int,
@@ -116,24 +116,24 @@ func newDKGProtocol(
round: round,
threshold: threshold,
sigToPub: sigToPub,
- idMap: make(map[types.ValidatorID]dkg.ID),
- mpkMap: make(map[types.ValidatorID]*dkg.PublicKeyShares),
+ idMap: make(map[types.NodeID]dkg.ID),
+ mpkMap: make(map[types.NodeID]*dkg.PublicKeyShares),
masterPrivateShare: prvShare,
prvShares: dkg.NewEmptyPrivateKeyShares(),
- prvSharesReceived: make(map[types.ValidatorID]struct{}),
+ prvSharesReceived: make(map[types.NodeID]struct{}),
}
}
func (d *dkgProtocol) processMasterPublicKeys(
mpks []*types.DKGMasterPublicKey) error {
- d.idMap = make(map[types.ValidatorID]dkg.ID, len(mpks))
- d.mpkMap = make(map[types.ValidatorID]*dkg.PublicKeyShares, len(mpks))
- d.prvSharesReceived = make(map[types.ValidatorID]struct{}, len(mpks))
+ d.idMap = make(map[types.NodeID]dkg.ID, len(mpks))
+ d.mpkMap = make(map[types.NodeID]*dkg.PublicKeyShares, len(mpks))
+ d.prvSharesReceived = make(map[types.NodeID]struct{}, len(mpks))
ids := make(dkg.IDs, len(mpks))
for i := range mpks {
- vID := mpks[i].ProposerID
- d.idMap[vID] = mpks[i].DKGID
- d.mpkMap[vID] = &mpks[i].PublicKeyShares
+ nID := mpks[i].ProposerID
+ d.idMap[nID] = mpks[i].DKGID
+ d.mpkMap[nID] = &mpks[i].PublicKeyShares
ids[i] = mpks[i].DKGID
}
d.masterPrivateShare.SetParticipants(ids)
@@ -152,15 +152,15 @@ func (d *dkgProtocol) processMasterPublicKeys(
}
func (d *dkgProtocol) proposeNackComplaints() {
- for vID := range d.mpkMap {
- if _, exist := d.prvSharesReceived[vID]; exist {
+ for nID := range d.mpkMap {
+ if _, exist := d.prvSharesReceived[nID]; exist {
continue
}
d.recv.ProposeDKGComplaint(&types.DKGComplaint{
ProposerID: d.ID,
Round: d.round,
PrivateShare: types.DKGPrivateShare{
- ProposerID: vID,
+ ProposerID: nID,
Round: d.round,
},
})
@@ -187,7 +187,7 @@ func (d *dkgProtocol) processPrivateShare(
return nil
}
self, exist := d.idMap[d.ID]
- // This validator is not a DKG participant, ignore the private share.
+ // This node is not a DKG participant, ignore the private share.
if !exist {
return nil
}
@@ -242,19 +242,19 @@ func newDKGGroupPublicKey(
threshold int, sigToPub SigToPubFn) (
*dkgGroupPublicKey, error) {
// Calculate qualify members.
- complaintsByID := map[types.ValidatorID]int{}
+ complaintsByID := map[types.NodeID]int{}
for _, complaint := range complaints {
complaintsByID[complaint.PrivateShare.ProposerID]++
}
- disqualifyIDs := map[types.ValidatorID]struct{}{}
- for vID, num := range complaintsByID {
+ disqualifyIDs := map[types.NodeID]struct{}{}
+ for nID, num := range complaintsByID {
if num > threshold {
- disqualifyIDs[vID] = struct{}{}
+ disqualifyIDs[nID] = struct{}{}
}
}
qualifyIDs := make(dkg.IDs, 0, len(mpks)-len(disqualifyIDs))
mpkMap := make(map[dkg.ID]*types.DKGMasterPublicKey, cap(qualifyIDs))
- idMap := make(map[types.ValidatorID]dkg.ID)
+ idMap := make(map[types.NodeID]dkg.ID)
for _, mpk := range mpks {
if _, exist := disqualifyIDs[mpk.ProposerID]; exist {
continue
@@ -264,7 +264,7 @@ func newDKGGroupPublicKey(
qualifyIDs = append(qualifyIDs, mpk.DKGID)
}
// Recover qualify members' public key.
- pubKeys := make(map[types.ValidatorID]*dkg.PublicKey, len(qualifyIDs))
+ pubKeys := make(map[types.NodeID]*dkg.PublicKey, len(qualifyIDs))
for _, recvID := range qualifyIDs {
pubShares := dkg.NewEmptyPublicKeyShares()
for _, id := range qualifyIDs {
diff --git a/core/dkg-tsig-protocol_test.go b/core/dkg-tsig-protocol_test.go
index 7da8866..fc64788 100644
--- a/core/dkg-tsig-protocol_test.go
+++ b/core/dkg-tsig-protocol_test.go
@@ -33,18 +33,18 @@ import (
type DKGTSIGProtocolTestSuite struct {
suite.Suite
- vIDs types.ValidatorIDs
- dkgIDs map[types.ValidatorID]dkg.ID
- prvKeys map[types.ValidatorID]crypto.PrivateKey
+ nIDs types.NodeIDs
+ dkgIDs map[types.NodeID]dkg.ID
+ prvKeys map[types.NodeID]crypto.PrivateKey
}
type testDKGReceiver struct {
s *DKGTSIGProtocolTestSuite
prvKey crypto.PrivateKey
- complaints map[types.ValidatorID]*types.DKGComplaint
+ complaints map[types.NodeID]*types.DKGComplaint
mpk *types.DKGMasterPublicKey
- prvShare map[types.ValidatorID]*types.DKGPrivateShare
+ prvShare map[types.NodeID]*types.DKGPrivateShare
}
func newTestDKGReceiver(
@@ -52,8 +52,8 @@ func newTestDKGReceiver(
return &testDKGReceiver{
s: s,
prvKey: prvKey,
- complaints: make(map[types.ValidatorID]*types.DKGComplaint),
- prvShare: make(map[types.ValidatorID]*types.DKGPrivateShare),
+ complaints: make(map[types.NodeID]*types.DKGComplaint),
+ prvShare: make(map[types.NodeID]*types.DKGPrivateShare),
}
}
@@ -72,7 +72,7 @@ func (r *testDKGReceiver) ProposeDKGMasterPublicKey(
r.mpk = mpk
}
func (r *testDKGReceiver) ProposeDKGPrivateShare(
- to types.ValidatorID, prv *types.DKGPrivateShare) {
+ to types.NodeID, prv *types.DKGPrivateShare) {
var err error
prv.Signature, err = r.prvKey.Sign(hashDKGPrivateShare(prv))
r.s.Require().NoError(err)
@@ -80,38 +80,38 @@ func (r *testDKGReceiver) ProposeDKGPrivateShare(
}
func (s *DKGTSIGProtocolTestSuite) setupDKGParticipants(n int) {
- s.vIDs = make(types.ValidatorIDs, 0, n)
- s.prvKeys = make(map[types.ValidatorID]crypto.PrivateKey, n)
- s.dkgIDs = make(map[types.ValidatorID]dkg.ID)
+ s.nIDs = make(types.NodeIDs, 0, n)
+ s.prvKeys = make(map[types.NodeID]crypto.PrivateKey, n)
+ s.dkgIDs = make(map[types.NodeID]dkg.ID)
ids := make(dkg.IDs, 0, n)
for i := 0; i < n; i++ {
prvKey, err := eth.NewPrivateKey()
s.Require().NoError(err)
- vID := types.NewValidatorID(prvKey.PublicKey())
- s.vIDs = append(s.vIDs, vID)
- s.prvKeys[vID] = prvKey
- id := dkg.NewID(vID.Hash[:])
+ nID := types.NewNodeID(prvKey.PublicKey())
+ s.nIDs = append(s.nIDs, nID)
+ s.prvKeys[nID] = prvKey
+ id := dkg.NewID(nID.Hash[:])
ids = append(ids, id)
- s.dkgIDs[vID] = id
+ s.dkgIDs[nID] = id
}
}
func (s *DKGTSIGProtocolTestSuite) newProtocols(k, n int, round uint64) (
- map[types.ValidatorID]*testDKGReceiver, map[types.ValidatorID]*dkgProtocol) {
+ map[types.NodeID]*testDKGReceiver, map[types.NodeID]*dkgProtocol) {
s.setupDKGParticipants(n)
- receivers := make(map[types.ValidatorID]*testDKGReceiver, n)
- protocols := make(map[types.ValidatorID]*dkgProtocol, n)
- for _, vID := range s.vIDs {
- receivers[vID] = newTestDKGReceiver(s, s.prvKeys[vID])
- protocols[vID] = newDKGProtocol(
- vID,
- receivers[vID],
+ receivers := make(map[types.NodeID]*testDKGReceiver, n)
+ protocols := make(map[types.NodeID]*dkgProtocol, n)
+ for _, nID := range s.nIDs {
+ receivers[nID] = newTestDKGReceiver(s, s.prvKeys[nID])
+ protocols[nID] = newDKGProtocol(
+ nID,
+ receivers[nID],
round,
k,
eth.SigToPub,
)
- s.Require().NotNil(receivers[vID].mpk)
+ s.Require().NotNil(receivers[nID].mpk)
}
return receivers, protocols
}
@@ -140,8 +140,8 @@ func (s *DKGTSIGProtocolTestSuite) TestDKGTSIGProtocol() {
for _, receiver := range receivers {
s.Require().Len(receiver.prvShare, n)
- for vID, prvShare := range receiver.prvShare {
- s.Require().NoError(protocols[vID].processPrivateShare(prvShare))
+ for nID, prvShare := range receiver.prvShare {
+ s.Require().NoError(protocols[nID].processPrivateShare(prvShare))
}
}
@@ -166,26 +166,26 @@ func (s *DKGTSIGProtocolTestSuite) TestDKGTSIGProtocol() {
}
shareSecrets := make(
- map[types.ValidatorID]*dkgShareSecret, len(qualifyIDs))
+ map[types.NodeID]*dkgShareSecret, len(qualifyIDs))
- for vID, protocol := range protocols {
- _, exist := qualifyIDs[s.dkgIDs[vID]]
+ for nID, protocol := range protocols {
+ _, exist := qualifyIDs[s.dkgIDs[nID]]
s.Require().True(exist)
var err error
- shareSecrets[vID], err = protocol.recoverShareSecret(gpk.qualifyIDs)
+ shareSecrets[nID], err = protocol.recoverShareSecret(gpk.qualifyIDs)
s.Require().NoError(err)
}
tsig := newTSigProtocol(gpk)
msgHash := crypto.Keccak256Hash([]byte("🏖🍹"))
- for vID, shareSecret := range shareSecrets {
+ for nID, shareSecret := range shareSecrets {
psig := &types.DKGPartialSignature{
- ProposerID: vID,
+ ProposerID: nID,
Round: round,
PartialSignature: shareSecret.sign(msgHash),
}
var err error
- psig.Signature, err = s.prvKeys[vID].Sign(hashDKGPartialSignature(psig))
+ psig.Signature, err = s.prvKeys[nID].Sign(hashDKGPartialSignature(psig))
s.Require().NoError(err)
s.Require().NoError(tsig.processPartialSignature(msgHash, psig))
if len(tsig.sigs) > k {
@@ -207,7 +207,7 @@ func (s *DKGTSIGProtocolTestSuite) TestNackComplaint() {
receivers, protocols := s.newProtocols(k, n, round)
- byzantineID := s.vIDs[0]
+ byzantineID := s.nIDs[0]
for _, receiver := range receivers {
gov.AddDKGMasterPublicKey(receiver.mpk)
@@ -223,8 +223,8 @@ func (s *DKGTSIGProtocolTestSuite) TestNackComplaint() {
if senderID == byzantineID {
continue
}
- for vID, prvShare := range receiver.prvShare {
- s.Require().NoError(protocols[vID].processPrivateShare(prvShare))
+ for nID, prvShare := range receiver.prvShare {
+ s.Require().NoError(protocols[nID].processPrivateShare(prvShare))
}
}
@@ -250,8 +250,8 @@ func (s *DKGTSIGProtocolTestSuite) TestComplaint() {
receivers, protocols := s.newProtocols(k, n, round)
- byzantineID := s.vIDs[0]
- targetID := s.vIDs[1]
+ byzantineID := s.nIDs[0]
+ targetID := s.nIDs[1]
receiver := receivers[targetID]
protocol := protocols[targetID]
@@ -266,7 +266,7 @@ func (s *DKGTSIGProtocolTestSuite) TestComplaint() {
// These messages are not valid.
err = protocol.processPrivateShare(&types.DKGPrivateShare{
- ProposerID: types.ValidatorID{Hash: common.NewRandomHash()},
+ ProposerID: types.NodeID{Hash: common.NewRandomHash()},
Round: round,
})
s.Error(ErrNotDKGParticipant, err)
@@ -296,7 +296,7 @@ func (s *DKGTSIGProtocolTestSuite) TestQualifyIDs() {
receivers, _ := s.newProtocols(k, n, round)
- byzantineID := s.vIDs[0]
+ byzantineID := s.nIDs[0]
for _, receiver := range receivers {
gov.AddDKGMasterPublicKey(receiver.mpk)
@@ -304,9 +304,9 @@ func (s *DKGTSIGProtocolTestSuite) TestQualifyIDs() {
complaints := make([]*types.DKGComplaint, k+1)
for i := range complaints {
- vID := s.vIDs[i]
+ nID := s.nIDs[i]
complaints[i] = &types.DKGComplaint{
- ProposerID: vID,
+ ProposerID: nID,
Round: round,
PrivateShare: types.DKGPrivateShare{
ProposerID: byzantineID,
@@ -344,7 +344,7 @@ func (s *DKGTSIGProtocolTestSuite) TestPartialSignature() {
receivers, protocols := s.newProtocols(k, n, round)
- byzantineID := s.vIDs[0]
+ byzantineID := s.nIDs[0]
for _, receiver := range receivers {
gov.AddDKGMasterPublicKey(receiver.mpk)
@@ -360,8 +360,8 @@ func (s *DKGTSIGProtocolTestSuite) TestPartialSignature() {
if senderID == byzantineID {
continue
}
- for vID, prvShare := range receiver.prvShare {
- s.Require().NoError(protocols[vID].processPrivateShare(prvShare))
+ for nID, prvShare := range receiver.prvShare {
+ s.Require().NoError(protocols[nID].processPrivateShare(prvShare))
}
}
@@ -389,38 +389,38 @@ func (s *DKGTSIGProtocolTestSuite) TestPartialSignature() {
}
shareSecrets := make(
- map[types.ValidatorID]*dkgShareSecret, len(qualifyIDs))
+ map[types.NodeID]*dkgShareSecret, len(qualifyIDs))
- for vID, protocol := range protocols {
- _, exist := qualifyIDs[s.dkgIDs[vID]]
- if vID == byzantineID {
+ for nID, protocol := range protocols {
+ _, exist := qualifyIDs[s.dkgIDs[nID]]
+ if nID == byzantineID {
exist = !exist
}
s.Require().True(exist)
var err error
- shareSecrets[vID], err = protocol.recoverShareSecret(gpk.qualifyIDs)
+ shareSecrets[nID], err = protocol.recoverShareSecret(gpk.qualifyIDs)
s.Require().NoError(err)
}
tsig := newTSigProtocol(gpk)
msgHash := crypto.Keccak256Hash([]byte("🏖🍹"))
- byzantineID2 := s.vIDs[1]
- for vID, shareSecret := range shareSecrets {
+ byzantineID2 := s.nIDs[1]
+ for nID, shareSecret := range shareSecrets {
psig := &types.DKGPartialSignature{
- ProposerID: vID,
+ ProposerID: nID,
Round: round,
PartialSignature: shareSecret.sign(msgHash),
}
- if vID == byzantineID2 {
+ if nID == byzantineID2 {
psig.PartialSignature[0]++
}
var err error
- psig.Signature, err = s.prvKeys[vID].Sign(hashDKGPartialSignature(psig))
+ psig.Signature, err = s.prvKeys[nID].Sign(hashDKGPartialSignature(psig))
s.Require().NoError(err)
err = tsig.processPartialSignature(msgHash, psig)
- if vID == byzantineID {
+ if nID == byzantineID {
s.Require().Error(ErrNotQualifyDKGParticipant, err)
- } else if vID == byzantineID2 {
+ } else if nID == byzantineID2 {
s.Require().Error(ErrIncorrectPartialSignature, err)
} else {
s.Require().NoError(err)
diff --git a/core/interfaces.go b/core/interfaces.go
index 3fae590..4f67e1e 100644
--- a/core/interfaces.go
+++ b/core/interfaces.go
@@ -63,7 +63,7 @@ type Network interface {
BroadcastWitnessAck(witnessAck *types.WitnessAck)
// SendDKGPrivateShare sends PrivateShare to a DKG participant.
- SendDKGPrivateShare(recv types.ValidatorID, prvShare *types.DKGPrivateShare)
+ SendDKGPrivateShare(recv types.NodeID, prvShare *types.DKGPrivateShare)
// ReceiveChan returns a channel to receive messages from DEXON network.
ReceiveChan() <-chan interface{}
@@ -76,8 +76,8 @@ type Governance interface {
// GetConfiguration returns the configuration at a given block height.
GetConfiguration(blockHeight uint64) *types.Config
- // Get the current validator set.
- GetNotarySet() map[types.ValidatorID]struct{}
+ // Get the current notary set.
+ GetNotarySet() map[types.NodeID]struct{}
//// DKG-related methods.
diff --git a/core/leader-selector_test.go b/core/leader-selector_test.go
index e18dd49..7eb9b3c 100644
--- a/core/leader-selector_test.go
+++ b/core/leader-selector_test.go
@@ -85,7 +85,7 @@ func (s *LeaderSelectorTestSuite) TestLeaderBlockHash() {
prv, err := eth.NewPrivateKey()
s.Require().Nil(err)
block := &types.Block{
- ProposerID: types.NewValidatorID(prv.PublicKey()),
+ ProposerID: types.NewNodeID(prv.PublicKey()),
Hash: common.NewRandomHash(),
}
s.Require().Nil(leader.prepareBlock(block, prv))
@@ -110,7 +110,7 @@ func (s *LeaderSelectorTestSuite) TestPrepareBlock() {
prv, err := eth.NewPrivateKey()
s.Require().Nil(err)
block := &types.Block{
- ProposerID: types.NewValidatorID(prv.PublicKey()),
+ ProposerID: types.NewNodeID(prv.PublicKey()),
}
s.Require().Nil(leader.prepareBlock(block, prv))
s.Nil(leader.processBlock(block))
diff --git a/core/negative-ack.go b/core/negative-ack.go
index 13a4832..8957152 100644
--- a/core/negative-ack.go
+++ b/core/negative-ack.go
@@ -25,67 +25,67 @@ import (
type negativeAck struct {
// owner is the ID of proposer itself, this is used when deciding
- // a validator to be restricted or not.
- owner types.ValidatorID
+ // a node to be restricted or not.
+ owner types.NodeID
- numOfValidators int
+ numOfNodes int
// timeDelay and timeExpire are for nack timeout.
timeDelay time.Duration
timeExpire time.Duration
- // restricteds stores validators which has been restricted and the time it's
+ // restricteds stores nodes which has been restricted and the time it's
// restricted.
- restricteds map[types.ValidatorID]time.Time
+ restricteds map[types.NodeID]time.Time
- // lastVotes and lockedVotes store the votes for nack. lastVotes[vid1][vid2]
- // and lockedVotes[vid1][vid2] both mean that vid2 votes vid1. The difference
- // is lockedVotes works only when vid1 is restricted, so that the votes are
+ // lastVotes and lockedVotes store the votes for nack. lastVotes[nid1][nid2]
+ // and lockedVotes[nid1][nid2] both mean that nid2 votes nid1. The difference
+ // is lockedVotes works only when nid1 is restricted, so that the votes are
// needed to be locked.
- lastVotes map[types.ValidatorID]map[types.ValidatorID]struct{}
- lockedVotes map[types.ValidatorID]map[types.ValidatorID]struct{}
+ lastVotes map[types.NodeID]map[types.NodeID]struct{}
+ lockedVotes map[types.NodeID]map[types.NodeID]struct{}
- // timeDiffs is the cache for last time stamps. timeDiffs[vid1][vid2] means
- // the last updated timestamps vid1 sees vid2.
- timeDiffs map[types.ValidatorID]map[types.ValidatorID]map[types.ValidatorID]time.Time
+ // timeDiffs is the cache for last time stamps. timeDiffs[nid1][nid2] means
+ // the last updated timestamps nid1 sees nid2.
+ timeDiffs map[types.NodeID]map[types.NodeID]map[types.NodeID]time.Time
}
// newNegativeAck creates a new negaticeAck instance.
-func newNegativeAck(vid types.ValidatorID) *negativeAck {
+func newNegativeAck(nid types.NodeID) *negativeAck {
n := &negativeAck{
- owner: vid,
- numOfValidators: 0,
- restricteds: make(map[types.ValidatorID]time.Time),
- lastVotes: make(map[types.ValidatorID]map[types.ValidatorID]struct{}),
- lockedVotes: make(map[types.ValidatorID]map[types.ValidatorID]struct{}),
- timeDiffs: make(map[types.ValidatorID]map[types.ValidatorID]map[types.ValidatorID]time.Time),
+ owner: nid,
+ numOfNodes: 0,
+ restricteds: make(map[types.NodeID]time.Time),
+ lastVotes: make(map[types.NodeID]map[types.NodeID]struct{}),
+ lockedVotes: make(map[types.NodeID]map[types.NodeID]struct{}),
+ timeDiffs: make(map[types.NodeID]map[types.NodeID]map[types.NodeID]time.Time),
}
- n.addValidator(vid)
+ n.addNode(nid)
return n
}
-// processNewVote is called when a new "vote" occurs, that is, a validator
-// sees that other 2f + 1 validators think a validator is slow. "vid" is the
-// validator which propesed the block which the timestamps votes and "h" is
-// the validator been voted to be nacked.
+// processNewVote is called when a new "vote" occurs, that is, a node
+// sees that other 2f + 1 nodes think a node is slow. "nid" is the
+// node which propesed the block which the timestamps votes and "h" is
+// the node been voted to be nacked.
func (n *negativeAck) processNewVote(
- vid types.ValidatorID,
- h types.ValidatorID,
-) []types.ValidatorID {
+ nid types.NodeID,
+ h types.NodeID,
+) []types.NodeID {
- nackeds := []types.ValidatorID{}
+ nackeds := []types.NodeID{}
if _, exist := n.restricteds[h]; exist {
- n.lockedVotes[h][vid] = struct{}{}
- if len(n.lockedVotes[h]) > 2*(n.numOfValidators-1)/3 {
+ n.lockedVotes[h][nid] = struct{}{}
+ if len(n.lockedVotes[h]) > 2*(n.numOfNodes-1)/3 {
nackeds = append(nackeds, h)
delete(n.restricteds, h)
}
} else {
- if n.owner == vid {
+ if n.owner == nid {
n.restrict(h)
} else {
- n.lastVotes[h][vid] = struct{}{}
- if len(n.lastVotes[h]) > (n.numOfValidators-1)/3 {
+ n.lastVotes[h][nid] = struct{}{}
+ if len(n.lastVotes[h]) > (n.numOfNodes-1)/3 {
n.restrict(h)
}
}
@@ -94,44 +94,44 @@ func (n *negativeAck) processNewVote(
}
// processTimestamps process new timestamps of a block which is proposed by
-// validator vid, and returns the validators being nacked.
+// node nid, and returns the nodes being nacked.
func (n *negativeAck) processTimestamps(
- vid types.ValidatorID,
- ts map[types.ValidatorID]time.Time,
-) []types.ValidatorID {
+ nid types.NodeID,
+ ts map[types.NodeID]time.Time,
+) []types.NodeID {
n.checkRestrictExpire()
- nackeds := []types.ValidatorID{}
+ nackeds := []types.NodeID{}
for h := range n.timeDiffs {
- if n.timeDiffs[vid][h][h].Equal(ts[h]) {
+ if n.timeDiffs[nid][h][h].Equal(ts[h]) {
votes := 0
for hh := range n.timeDiffs {
- if ts[hh].Sub(n.timeDiffs[vid][h][hh]) >= n.timeDelay {
+ if ts[hh].Sub(n.timeDiffs[nid][h][hh]) >= n.timeDelay {
votes++
}
}
- if votes > 2*((n.numOfValidators-1)/3) {
- n.lastVotes[h][vid] = struct{}{}
- nack := n.processNewVote(vid, h)
+ if votes > 2*((n.numOfNodes-1)/3) {
+ n.lastVotes[h][nid] = struct{}{}
+ nack := n.processNewVote(nid, h)
for _, i := range nack {
nackeds = append(nackeds, i)
}
} else {
- delete(n.lastVotes[h], vid)
+ delete(n.lastVotes[h], nid)
}
} else {
for hh := range n.timeDiffs {
- n.timeDiffs[vid][h][hh] = ts[hh]
+ n.timeDiffs[nid][h][hh] = ts[hh]
}
- delete(n.lastVotes[h], vid)
+ delete(n.lastVotes[h], nid)
}
}
return nackeds
}
func (n *negativeAck) checkRestrictExpire() {
- expired := []types.ValidatorID{}
+ expired := []types.NodeID{}
now := time.Now()
for h, t := range n.restricteds {
if now.Sub(t) >= n.timeExpire {
@@ -143,19 +143,19 @@ func (n *negativeAck) checkRestrictExpire() {
}
}
-func (n *negativeAck) restrict(vid types.ValidatorID) {
- if _, exist := n.restricteds[vid]; !exist {
- n.restricteds[vid] = time.Now().UTC()
- n.lockedVotes[vid] = map[types.ValidatorID]struct{}{}
- for h := range n.lastVotes[vid] {
- n.lockedVotes[vid][h] = struct{}{}
+func (n *negativeAck) restrict(nid types.NodeID) {
+ if _, exist := n.restricteds[nid]; !exist {
+ n.restricteds[nid] = time.Now().UTC()
+ n.lockedVotes[nid] = map[types.NodeID]struct{}{}
+ for h := range n.lastVotes[nid] {
+ n.lockedVotes[nid][h] = struct{}{}
}
}
}
-func (n *negativeAck) getRestrictedValidators() map[types.ValidatorID]struct{} {
+func (n *negativeAck) getRestrictedNodes() map[types.NodeID]struct{} {
n.checkRestrictExpire()
- ret := map[types.ValidatorID]struct{}{}
+ ret := map[types.NodeID]struct{}{}
for h := range n.restricteds {
ret[h] = struct{}{}
}
@@ -170,42 +170,42 @@ func (n *negativeAck) setTimeExpire(t time.Duration) {
n.timeExpire = t
}
-func (n *negativeAck) addValidator(vid types.ValidatorID) {
- n.numOfValidators++
- n.lastVotes[vid] = make(map[types.ValidatorID]struct{})
- n.lockedVotes[vid] = make(map[types.ValidatorID]struct{})
+func (n *negativeAck) addNode(nid types.NodeID) {
+ n.numOfNodes++
+ n.lastVotes[nid] = make(map[types.NodeID]struct{})
+ n.lockedVotes[nid] = make(map[types.NodeID]struct{})
- newTimeDiff := make(map[types.ValidatorID]map[types.ValidatorID]time.Time)
+ newTimeDiff := make(map[types.NodeID]map[types.NodeID]time.Time)
for h := range n.timeDiffs {
- newTimeDiff2 := make(map[types.ValidatorID]time.Time)
+ newTimeDiff2 := make(map[types.NodeID]time.Time)
for hh := range n.timeDiffs {
newTimeDiff2[hh] = time.Time{}
}
newTimeDiff[h] = newTimeDiff2
}
- n.timeDiffs[vid] = newTimeDiff
+ n.timeDiffs[nid] = newTimeDiff
for h := range n.timeDiffs {
- n.timeDiffs[h][vid] = make(map[types.ValidatorID]time.Time)
+ n.timeDiffs[h][nid] = make(map[types.NodeID]time.Time)
}
}
-func (n *negativeAck) deleteValidator(vid types.ValidatorID) {
- n.numOfValidators--
+func (n *negativeAck) deleteNode(nid types.NodeID) {
+ n.numOfNodes--
- delete(n.timeDiffs, vid)
+ delete(n.timeDiffs, nid)
for h := range n.lastVotes {
- delete(n.lastVotes[h], vid)
+ delete(n.lastVotes[h], nid)
}
- delete(n.lastVotes, vid)
- delete(n.lockedVotes, vid)
+ delete(n.lastVotes, nid)
+ delete(n.lockedVotes, nid)
for h := range n.timeDiffs {
- delete(n.timeDiffs[h], vid)
+ delete(n.timeDiffs[h], nid)
for hh := range n.timeDiffs[h] {
- delete(n.timeDiffs[h][hh], vid)
+ delete(n.timeDiffs[h][hh], nid)
}
}
- delete(n.restricteds, vid)
+ delete(n.restricteds, nid)
}
diff --git a/core/negative-ack_test.go b/core/negative-ack_test.go
index 990871e..22280da 100644
--- a/core/negative-ack_test.go
+++ b/core/negative-ack_test.go
@@ -46,93 +46,93 @@ func (s *NegativeAckTest) SetupTest() {
}
func (s *NegativeAckTest) checkLastVotes(
- vids []types.ValidatorID,
- vs map[types.ValidatorID]map[types.ValidatorID]struct{},
+ nids []types.NodeID,
+ vs map[types.NodeID]map[types.NodeID]struct{},
a [][]bool,
) {
- for i := 0; i < len(vids); i++ {
- for j := 0; j < len(vids); j++ {
- _, exist := vs[vids[i]][vids[j]]
+ for i := 0; i < len(nids); i++ {
+ for j := 0; j < len(nids); j++ {
+ _, exist := vs[nids[i]][nids[j]]
s.Require().Equal(a[i][j], exist)
}
}
}
func (s *NegativeAckTest) checkTimeDiff(
- vids []types.ValidatorID,
- ts map[types.ValidatorID]map[types.ValidatorID]time.Time,
+ nids []types.NodeID,
+ ts map[types.NodeID]map[types.NodeID]time.Time,
a [][]int,
) {
- for i := 0; i < len(vids); i++ {
- for j := 0; j < len(vids); j++ {
+ for i := 0; i < len(nids); i++ {
+ for j := 0; j < len(nids); j++ {
s.Require().Equal(
time.Duration(a[i][j])*timeDelay,
- ts[vids[i]][vids[j]].Sub(baseTime),
+ ts[nids[i]][nids[j]].Sub(baseTime),
)
}
}
}
-func genTimestamp(vids []types.ValidatorID, a []int) map[types.ValidatorID]time.Time {
- ts := map[types.ValidatorID]time.Time{}
- for i := 0; i < len(vids); i++ {
- ts[vids[i]] = baseTime.Add(time.Duration(a[i]) * timeDelay)
+func genTimestamp(nids []types.NodeID, a []int) map[types.NodeID]time.Time {
+ ts := map[types.NodeID]time.Time{}
+ for i := 0; i < len(nids); i++ {
+ ts[nids[i]] = baseTime.Add(time.Duration(a[i]) * timeDelay)
}
return ts
}
-func genTestNegativeAck(num int) (*negativeAck, []types.ValidatorID) {
- vids := test.GenerateRandomValidatorIDs(num)
- n := newNegativeAck(vids[0])
+func genTestNegativeAck(num int) (*negativeAck, []types.NodeID) {
+ nids := test.GenerateRandomNodeIDs(num)
+ n := newNegativeAck(nids[0])
for i := 1; i < num; i++ {
- n.addValidator(vids[i])
+ n.addNode(nids[i])
}
- return n, vids
+ return n, nids
}
func (s *NegativeAckTest) TestProcessTimestamps() {
- n, vids := genTestNegativeAck(4)
+ n, nids := genTestNegativeAck(4)
n.setTimeDelay(timeDelay)
n.setTimeExpire(timeExpire)
- n.processTimestamps(vids[0], genTimestamp(vids, []int{1, 1, 1, 0}))
- s.checkTimeDiff(vids, n.timeDiffs[vids[0]], [][]int{
+ n.processTimestamps(nids[0], genTimestamp(nids, []int{1, 1, 1, 0}))
+ s.checkTimeDiff(nids, n.timeDiffs[nids[0]], [][]int{
{1, 1, 1, 0},
{1, 1, 1, 0},
{1, 1, 1, 0},
{1, 1, 1, 0},
})
- s.checkLastVotes(vids, n.lastVotes, [][]bool{
+ s.checkLastVotes(nids, n.lastVotes, [][]bool{
{false, false, false, false},
{false, false, false, false},
{false, false, false, false},
{false, false, false, false},
})
- n.processTimestamps(vids[0], genTimestamp(vids, []int{3, 1, 2, 1}))
- s.checkTimeDiff(vids, n.timeDiffs[vids[0]], [][]int{
+ n.processTimestamps(nids[0], genTimestamp(nids, []int{3, 1, 2, 1}))
+ s.checkTimeDiff(nids, n.timeDiffs[nids[0]], [][]int{
{3, 1, 2, 1},
{1, 1, 1, 0},
{3, 1, 2, 1},
{3, 1, 2, 1},
})
- s.checkLastVotes(vids, n.lastVotes, [][]bool{
+ s.checkLastVotes(nids, n.lastVotes, [][]bool{
{false, false, false, false},
{true, false, false, false},
{false, false, false, false},
{false, false, false, false},
})
- n.processTimestamps(vids[0], genTimestamp(vids, []int{5, 1, 2, 2}))
- s.checkTimeDiff(vids, n.timeDiffs[vids[0]], [][]int{
+ n.processTimestamps(nids[0], genTimestamp(nids, []int{5, 1, 2, 2}))
+ s.checkTimeDiff(nids, n.timeDiffs[nids[0]], [][]int{
{5, 1, 2, 2},
{1, 1, 1, 0},
{3, 1, 2, 1},
{5, 1, 2, 2},
})
- s.checkLastVotes(vids, n.lastVotes, [][]bool{
+ s.checkLastVotes(nids, n.lastVotes, [][]bool{
{false, false, false, false},
{true, false, false, false},
{false, false, false, false},
@@ -142,83 +142,83 @@ func (s *NegativeAckTest) TestProcessTimestamps() {
func (s *NegativeAckTest) TestRestrictBySelf() {
var exist bool
- n, vids := genTestNegativeAck(4)
+ n, nids := genTestNegativeAck(4)
n.setTimeDelay(timeDelay)
n.setTimeExpire(timeExpire)
- n.processTimestamps(vids[0], genTimestamp(vids, []int{1, 1, 1, 0}))
- _, exist = n.getRestrictedValidators()[vids[1]]
+ n.processTimestamps(nids[0], genTimestamp(nids, []int{1, 1, 1, 0}))
+ _, exist = n.getRestrictedNodes()[nids[1]]
s.Require().False(exist)
- n.processTimestamps(vids[0], genTimestamp(vids, []int{3, 1, 2, 1}))
- _, exist = n.getRestrictedValidators()[vids[1]]
+ n.processTimestamps(nids[0], genTimestamp(nids, []int{3, 1, 2, 1}))
+ _, exist = n.getRestrictedNodes()[nids[1]]
s.Require().True(exist)
}
func (s *NegativeAckTest) TestRestrictByVoting() {
- var nackeds []types.ValidatorID
+ var nackeds []types.NodeID
var exist bool
- n, vids := genTestNegativeAck(4)
+ n, nids := genTestNegativeAck(4)
n.setTimeDelay(timeDelay)
n.setTimeExpire(timeExpire)
- n.processTimestamps(vids[0], genTimestamp(vids, []int{1, 1, 1, 1}))
- n.processTimestamps(vids[0], genTimestamp(vids, []int{2, 2, 2, 2}))
+ n.processTimestamps(nids[0], genTimestamp(nids, []int{1, 1, 1, 1}))
+ n.processTimestamps(nids[0], genTimestamp(nids, []int{2, 2, 2, 2}))
- n.processTimestamps(vids[1], genTimestamp(vids, []int{1, 1, 1, 1}))
- n.processTimestamps(vids[2], genTimestamp(vids, []int{1, 1, 1, 1}))
- n.processTimestamps(vids[3], genTimestamp(vids, []int{1, 1, 1, 1}))
+ n.processTimestamps(nids[1], genTimestamp(nids, []int{1, 1, 1, 1}))
+ n.processTimestamps(nids[2], genTimestamp(nids, []int{1, 1, 1, 1}))
+ n.processTimestamps(nids[3], genTimestamp(nids, []int{1, 1, 1, 1}))
- nackeds = n.processTimestamps(vids[1], genTimestamp(vids, []int{1, 3, 3, 3}))
- _, exist = n.getRestrictedValidators()[vids[0]]
+ nackeds = n.processTimestamps(nids[1], genTimestamp(nids, []int{1, 3, 3, 3}))
+ _, exist = n.getRestrictedNodes()[nids[0]]
s.Require().False(exist)
s.Require().Equal(0, len(nackeds))
- nackeds = n.processTimestamps(vids[2], genTimestamp(vids, []int{1, 3, 3, 3}))
- _, exist = n.getRestrictedValidators()[vids[0]]
+ nackeds = n.processTimestamps(nids[2], genTimestamp(nids, []int{1, 3, 3, 3}))
+ _, exist = n.getRestrictedNodes()[nids[0]]
s.Require().True(exist)
s.Require().Equal(0, len(nackeds))
- nackeds = n.processTimestamps(vids[3], genTimestamp(vids, []int{1, 3, 3, 3}))
- _, exist = n.getRestrictedValidators()[vids[0]]
+ nackeds = n.processTimestamps(nids[3], genTimestamp(nids, []int{1, 3, 3, 3}))
+ _, exist = n.getRestrictedNodes()[nids[0]]
s.Require().False(exist)
s.Require().Equal(1, len(nackeds))
- s.Require().Equal(vids[0], nackeds[0])
+ s.Require().Equal(nids[0], nackeds[0])
}
func (s *NegativeAckTest) TestExpire() {
var exist bool
- n, vids := genTestNegativeAck(4)
+ n, nids := genTestNegativeAck(4)
n.setTimeDelay(timeDelay)
n.setTimeExpire(timeExpire)
- n.processTimestamps(vids[0], genTimestamp(vids, []int{1, 1, 1, 1}))
- n.processTimestamps(vids[1], genTimestamp(vids, []int{1, 1, 1, 1}))
- n.processTimestamps(vids[2], genTimestamp(vids, []int{1, 1, 1, 1}))
- n.processTimestamps(vids[3], genTimestamp(vids, []int{1, 1, 1, 1}))
+ n.processTimestamps(nids[0], genTimestamp(nids, []int{1, 1, 1, 1}))
+ n.processTimestamps(nids[1], genTimestamp(nids, []int{1, 1, 1, 1}))
+ n.processTimestamps(nids[2], genTimestamp(nids, []int{1, 1, 1, 1}))
+ n.processTimestamps(nids[3], genTimestamp(nids, []int{1, 1, 1, 1}))
- n.processTimestamps(vids[1], genTimestamp(vids, []int{1, 3, 3, 3}))
- n.processTimestamps(vids[2], genTimestamp(vids, []int{1, 3, 3, 3}))
- _, exist = n.getRestrictedValidators()[vids[0]]
+ n.processTimestamps(nids[1], genTimestamp(nids, []int{1, 3, 3, 3}))
+ n.processTimestamps(nids[2], genTimestamp(nids, []int{1, 3, 3, 3}))
+ _, exist = n.getRestrictedNodes()[nids[0]]
s.Require().True(exist)
time.Sleep(2 * timeExpire)
- n.processTimestamps(vids[0], genTimestamp(vids, []int{2, 2, 2, 2}))
+ n.processTimestamps(nids[0], genTimestamp(nids, []int{2, 2, 2, 2}))
- _, exist = n.getRestrictedValidators()[vids[0]]
+ _, exist = n.getRestrictedNodes()[nids[0]]
s.Require().False(exist)
}
-func (s *NegativeAckTest) TestAddDeleteValidator() {
- n, vids := genTestNegativeAck(10)
+func (s *NegativeAckTest) TestAddDeleteNode() {
+ n, nids := genTestNegativeAck(10)
s.Require().Equal(10, len(n.timeDiffs))
- s.Require().Equal(10, len(n.timeDiffs[vids[0]]))
+ s.Require().Equal(10, len(n.timeDiffs[nids[0]]))
- for _, vid := range vids {
- n.deleteValidator(vid)
+ for _, nid := range nids {
+ n.deleteNode(nid)
}
s.Require().Equal(0, len(n.timeDiffs))
}
diff --git a/core/reliable-broadcast.go b/core/reliable-broadcast.go
index 1681270..3c290ff 100644
--- a/core/reliable-broadcast.go
+++ b/core/reliable-broadcast.go
@@ -38,8 +38,8 @@ const (
// reliableBroadcast is a module for reliable broadcast.
type reliableBroadcast struct {
- // lattice stores validator's blocks and other info.
- lattice []*rbcValidatorStatus
+ // lattice stores node's blocks and other info.
+ lattice []*rbcNodeStatus
// blockInfos stores block infos.
blockInfos map[common.Hash]*rbcBlockInfo
@@ -48,12 +48,12 @@ type reliableBroadcast struct {
// in lattice.
receivedBlocks map[common.Hash]*types.Block
- // validators stores validator set.
- validators map[types.ValidatorID]struct{}
+ // nodes stores node set.
+ nodes map[types.NodeID]struct{}
}
-type rbcValidatorStatus struct {
- // blocks stores blocks proposed by specified validator in map which key is
+type rbcNodeStatus struct {
+ // blocks stores blocks proposed by specified node in map which key is
// the height of the block.
blocks map[uint64]*types.Block
@@ -94,7 +94,7 @@ func newReliableBroadcast() *reliableBroadcast {
return &reliableBroadcast{
blockInfos: make(map[common.Hash]*rbcBlockInfo),
receivedBlocks: make(map[common.Hash]*types.Block),
- validators: make(map[types.ValidatorID]struct{}),
+ nodes: make(map[types.NodeID]struct{}),
}
}
@@ -104,8 +104,8 @@ func (rb *reliableBroadcast) sanityCheck(b *types.Block) error {
return ErrInvalidChainID
}
- // Check if its proposer is in validator set.
- if _, exist := rb.validators[b.ProposerID]; !exist {
+ // Check if its proposer is in node set.
+ if _, exist := rb.nodes[b.ProposerID]; !exist {
return ErrInvalidProposerID
}
@@ -224,7 +224,7 @@ func (rb *reliableBroadcast) processBlock(block *types.Block) (err error) {
for _, h := range b.Acks {
bAckStat := rb.blockInfos[h]
// Update nextAck only when bAckStat.block.Position.Height + 1
- // is greater. A block might ack blocks proposed by same validator with
+ // is greater. A block might ack blocks proposed by same node with
// different height.
if rb.lattice[chainID].nextAck[bAckStat.block.Position.ChainID] <
bAckStat.block.Position.Height+1 {
@@ -241,7 +241,7 @@ func (rb *reliableBroadcast) processBlock(block *types.Block) (err error) {
}
bAckStat.ackedChain[chainID] = struct{}{}
// A block is strongly acked if it is acked by more than
- // 2 * (maximum number of byzatine validators) unique validators.
+ // 2 * (maximum number of byzatine nodes) unique nodes.
if len(bAckStat.ackedChain) > 2*((len(rb.lattice)-1)/3) {
blocksToAcked[bAckStat.block.Hash] = bAckStat.block
}
@@ -272,8 +272,8 @@ func (rb *reliableBroadcast) processBlock(block *types.Block) (err error) {
// Delete old blocks in "lattice" and "blocks" for release memory space.
// First, find the height that blocks below it can be deleted. This height
- // is defined by finding minimum of validator's nextOutput and last acking
- // heights from other validators, i.e. rb.lattice[v_other].nextAck[this_vid].
+ // is defined by finding minimum of node's nextOutput and last acking
+ // heights from other nodes, i.e. rb.lattice[v_other].nextAck[this_vid].
// This works because blocks of height below this minimum are not going to be
// acked anymore, the ackings of these blocks are illegal.
for vid := range rb.lattice {
@@ -351,7 +351,7 @@ func (rb *reliableBroadcast) extractBlocks() []*types.Block {
// prepareBlock helps to setup fields of block based on its ProposerID,
// including:
-// - Set 'Acks' and 'Timestamps' for the highest block of each validator not
+// - Set 'Acks' and 'Timestamps' for the highest block of each node not
// acked by this proposer before.
// - Set 'ParentHash' and 'Height' from parent block, if we can't find a
// parent, these fields would be setup like a genesis block.
@@ -361,7 +361,7 @@ func (rb *reliableBroadcast) prepareBlock(block *types.Block) {
block.ParentHash = common.Hash{}
acks := common.Hashes{}
for chainID := range rb.lattice {
- // find height of the latest block for that validator.
+ // find height of the latest block for that node.
var (
curBlock *types.Block
nextHeight = rb.lattice[block.Position.ChainID].nextAck[chainID]
@@ -394,21 +394,21 @@ func (rb *reliableBroadcast) prepareBlock(block *types.Block) {
return
}
-// addValidator adds validator in the validator set.
-func (rb *reliableBroadcast) addValidator(h types.ValidatorID) {
- rb.validators[h] = struct{}{}
+// addNode adds node in the node set.
+func (rb *reliableBroadcast) addNode(h types.NodeID) {
+ rb.nodes[h] = struct{}{}
}
-// deleteValidator deletes validator in validator set.
-func (rb *reliableBroadcast) deleteValidator(h types.ValidatorID) {
- delete(rb.validators, h)
+// deleteNode deletes node in node set.
+func (rb *reliableBroadcast) deleteNode(h types.NodeID) {
+ delete(rb.nodes, h)
}
// setChainNum set the number of chains.
func (rb *reliableBroadcast) setChainNum(num uint32) {
- rb.lattice = make([]*rbcValidatorStatus, num)
+ rb.lattice = make([]*rbcNodeStatus, num)
for i := range rb.lattice {
- rb.lattice[i] = &rbcValidatorStatus{
+ rb.lattice[i] = &rbcNodeStatus{
blocks: make(map[uint64]*types.Block),
nextAck: make([]uint64, num),
nextOutput: 0,
diff --git a/core/reliable-broadcast_test.go b/core/reliable-broadcast_test.go
index b1fc672..cf82949 100644
--- a/core/reliable-broadcast_test.go
+++ b/core/reliable-broadcast_test.go
@@ -46,8 +46,8 @@ func (s *ReliableBroadcastTest) SetupTest() {
}
func (s *ReliableBroadcastTest) prepareGenesisBlock(
- proposerID types.ValidatorID,
- validatorIDs []types.ValidatorID) (b *types.Block) {
+ proposerID types.NodeID,
+ nodeIDs []types.NodeID) (b *types.Block) {
b = &types.Block{
ProposerID: proposerID,
@@ -58,7 +58,7 @@ func (s *ReliableBroadcastTest) prepareGenesisBlock(
Acks: common.NewSortedHashes(common.Hashes{}),
Timestamp: time.Now().UTC(),
}
- for i, vID := range validatorIDs {
+ for i, vID := range nodeIDs {
if proposerID == vID {
b.Position.ChainID = uint32(i)
break
@@ -79,16 +79,16 @@ func (s *ReliableBroadcastTest) prepareGenesisBlock(
// 1 | 1
// | | |
// 0 0 0 0 (block height)
-// 0 1 2 3 (validator)
-func genTestCase1(s *ReliableBroadcastTest, rb *reliableBroadcast) []types.ValidatorID {
- // Create new reliableBroadcast instance with 4 validators
+// 0 1 2 3 (node)
+func genTestCase1(s *ReliableBroadcastTest, rb *reliableBroadcast) []types.NodeID {
+ // Create new reliableBroadcast instance with 4 nodes
var b *types.Block
var h common.Hash
- vids := []types.ValidatorID{}
+ vids := []types.NodeID{}
for i := 0; i < 4; i++ {
- vid := types.ValidatorID{Hash: common.NewRandomHash()}
- rb.addValidator(vid)
+ vid := types.NodeID{Hash: common.NewRandomHash()}
+ rb.addNode(vid)
vids = append(vids, vid)
}
rb.setChainNum(uint32(len(vids)))
@@ -176,20 +176,20 @@ func genTestCase1(s *ReliableBroadcastTest, rb *reliableBroadcast) []types.Valid
return vids
}
-func (s *ReliableBroadcastTest) TestAddValidator() {
+func (s *ReliableBroadcastTest) TestAddNode() {
rb := newReliableBroadcast()
s.Require().Equal(len(rb.lattice), 0)
vids := genTestCase1(s, rb)
s.Require().Equal(len(rb.lattice), 4)
for _, vid := range vids {
- rb.deleteValidator(vid)
+ rb.deleteNode(vid)
}
}
func (s *ReliableBroadcastTest) TestSanityCheck() {
var b *types.Block
var h common.Hash
- var vids []types.ValidatorID
+ var vids []types.NodeID
var err error
rb := newReliableBroadcast()
vids = genTestCase1(s, rb)
@@ -247,7 +247,7 @@ func (s *ReliableBroadcastTest) TestSanityCheck() {
// Invalid proposer ID.
h = rb.lattice[1].blocks[0].Hash
b = &types.Block{
- ProposerID: types.ValidatorID{Hash: common.NewRandomHash()},
+ ProposerID: types.NodeID{Hash: common.NewRandomHash()},
ParentHash: h,
Position: types.Position{
Height: 1,
@@ -365,7 +365,7 @@ func (s *ReliableBroadcastTest) TestAreAllAcksInLattice() {
func (s *ReliableBroadcastTest) TestStrongAck() {
var b *types.Block
- var vids []types.ValidatorID
+ var vids []types.NodeID
rb := newReliableBroadcast()
vids = genTestCase1(s, rb)
@@ -491,13 +491,13 @@ func (s *ReliableBroadcastTest) TestExtractBlocks() {
func (s *ReliableBroadcastTest) TestRandomIntensiveAcking() {
rb := newReliableBroadcast()
- vids := test.GenerateRandomValidatorIDs(4)
- heights := map[types.ValidatorID]uint64{}
+ vids := test.GenerateRandomNodeIDs(4)
+ heights := map[types.NodeID]uint64{}
extractedBlocks := []*types.Block{}
- // Generate validators.
+ // Generate nodes.
for _, vid := range vids {
- rb.addValidator(vid)
+ rb.addNode(vid)
}
rb.setChainNum(uint32(len(vids)))
// Generate genesis blocks.
@@ -545,9 +545,9 @@ func (s *ReliableBroadcastTest) TestRandomIntensiveAcking() {
func (s *ReliableBroadcastTest) TestRandomlyGeneratedBlocks() {
var (
- validatorCount = 19
- blockCount = 50
- repeat = 20
+ nodeCount = 19
+ blockCount = 50
+ repeat = 20
)
// Prepare a randomly generated blocks.
@@ -560,7 +560,7 @@ func (s *ReliableBroadcastTest) TestRandomlyGeneratedBlocks() {
}
}()
gen := test.NewBlocksGenerator(nil, hashBlock)
- _, err = gen.Generate(validatorCount, blockCount, nil, db)
+ _, err = gen.Generate(nodeCount, blockCount, nil, db)
s.Require().Nil(err)
iter, err := db.GetAll()
s.Require().Nil(err)
@@ -570,9 +570,9 @@ func (s *ReliableBroadcastTest) TestRandomlyGeneratedBlocks() {
stronglyAckedHashesAsString := map[string]struct{}{}
for i := 0; i < repeat; i++ {
- validators := map[types.ValidatorID]struct{}{}
+ nodes := map[types.NodeID]struct{}{}
rb := newReliableBroadcast()
- rb.setChainNum(uint32(validatorCount))
+ rb.setChainNum(uint32(nodeCount))
stronglyAckedHashes := common.Hashes{}
revealer.Reset()
@@ -587,10 +587,10 @@ func (s *ReliableBroadcastTest) TestRandomlyGeneratedBlocks() {
}
s.Require().Nil(err)
- // It's a hack to add validator to reliableBroadcast module.
- if _, added := validators[b.ProposerID]; !added {
- rb.addValidator(b.ProposerID)
- validators[b.ProposerID] = struct{}{}
+ // It's a hack to add node to reliableBroadcast module.
+ if _, added := nodes[b.ProposerID]; !added {
+ rb.addNode(b.ProposerID)
+ nodes[b.ProposerID] = struct{}{}
}
// Perform reliable broadcast process.
s.Require().Nil(rb.processBlock(&b))
@@ -621,21 +621,21 @@ func (s *ReliableBroadcastTest) TestPrepareBlock() {
req = s.Require()
rb = newReliableBroadcast()
minInterval = 50 * time.Millisecond
- validators = test.GenerateRandomValidatorIDs(4)
+ nodes = test.GenerateRandomNodeIDs(4)
)
- // Prepare validator IDs.
- for _, vID := range validators {
- rb.addValidator(vID)
+ // Prepare node IDs.
+ for _, vID := range nodes {
+ rb.addNode(vID)
}
- rb.setChainNum(uint32(len(validators)))
+ rb.setChainNum(uint32(len(nodes)))
// Setup genesis blocks.
- b00 := s.prepareGenesisBlock(validators[0], validators)
+ b00 := s.prepareGenesisBlock(nodes[0], nodes)
time.Sleep(minInterval)
- b10 := s.prepareGenesisBlock(validators[1], validators)
+ b10 := s.prepareGenesisBlock(nodes[1], nodes)
time.Sleep(minInterval)
- b20 := s.prepareGenesisBlock(validators[2], validators)
+ b20 := s.prepareGenesisBlock(nodes[2], nodes)
time.Sleep(minInterval)
- b30 := s.prepareGenesisBlock(validators[3], validators)
+ b30 := s.prepareGenesisBlock(nodes[3], nodes)
// Submit these blocks to reliableBroadcast instance.
s.Require().Nil(rb.processBlock(b00))
s.Require().Nil(rb.processBlock(b10))
@@ -644,7 +644,7 @@ func (s *ReliableBroadcastTest) TestPrepareBlock() {
// We should be able to collect all 4 genesis blocks by calling
// prepareBlock.
b11 := &types.Block{
- ProposerID: validators[1],
+ ProposerID: nodes[1],
Position: types.Position{
ChainID: 1,
},
@@ -664,7 +664,7 @@ func (s *ReliableBroadcastTest) TestPrepareBlock() {
s.Require().Nil(rb.processBlock(b11))
// Propose/Process a block based on collected info.
b12 := &types.Block{
- ProposerID: validators[1],
+ ProposerID: nodes[1],
Position: types.Position{
ChainID: 1,
},
@@ -677,10 +677,10 @@ func (s *ReliableBroadcastTest) TestPrepareBlock() {
req.Contains(b12.Acks, b11.Hash)
req.Equal(b12.ParentHash, b11.Hash)
req.Equal(b12.Position.Height, uint64(2))
- // When calling with other validator ID, we should be able to
+ // When calling with other node ID, we should be able to
// get 4 blocks to ack.
b01 := &types.Block{
- ProposerID: validators[0],
+ ProposerID: nodes[0],
Position: types.Position{
ChainID: 0,
},
diff --git a/core/test/blocks-generator.go b/core/test/blocks-generator.go
index 93867f1..759e3a6 100644
--- a/core/test/blocks-generator.go
+++ b/core/test/blocks-generator.go
@@ -28,94 +28,97 @@ import (
"github.com/dexon-foundation/dexon-consensus-core/core/types"
)
+// TODO(mission): blocks generator should generate blocks based on chain,
+// not nodes.
+
// ErrParentNotAcked would be raised when some block doesn't
// ack its parent block.
var ErrParentNotAcked = errors.New("parent is not acked")
-// validatorStatus is a state holder for each validator
+// nodeStatus is a state holder for each node
// during generating blocks.
-type validatorStatus struct {
+type nodeStatus struct {
blocks []*types.Block
- lastAckingHeight map[types.ValidatorID]uint64
+ lastAckingHeight map[types.NodeID]uint64
}
type hashBlockFn func(*types.Block) (common.Hash, error)
// getAckedBlockHash would randomly pick one block between
// last acked one to current head.
-func (vs *validatorStatus) getAckedBlockHash(
- ackedVID types.ValidatorID,
- ackedValidator *validatorStatus,
+func (vs *nodeStatus) getAckedBlockHash(
+ ackedNID types.NodeID,
+ ackedNode *nodeStatus,
randGen *rand.Rand) (
hash common.Hash, ok bool) {
- baseAckingHeight, exists := vs.lastAckingHeight[ackedVID]
+ baseAckingHeight, exists := vs.lastAckingHeight[ackedNID]
if exists {
// Do not ack the same block(height) twice.
baseAckingHeight++
}
- totalBlockCount := uint64(len(ackedValidator.blocks))
+ totalBlockCount := uint64(len(ackedNode.blocks))
if totalBlockCount <= baseAckingHeight {
// There is no new block to ack.
return
}
ackableRange := totalBlockCount - baseAckingHeight
height := uint64((randGen.Uint64() % ackableRange) + baseAckingHeight)
- vs.lastAckingHeight[ackedVID] = height
- hash = ackedValidator.blocks[height].Hash
+ vs.lastAckingHeight[ackedNID] = height
+ hash = ackedNode.blocks[height].Hash
ok = true
return
}
-// validatorSetStatus is a state holder for all validators
+// nodeSetStatus is a state holder for all nodes
// during generating blocks.
-type validatorSetStatus struct {
- status map[types.ValidatorID]*validatorStatus
- proposerChain map[types.ValidatorID]uint32
+type nodeSetStatus struct {
+ status map[types.NodeID]*nodeStatus
+ proposerChain map[types.NodeID]uint32
timestamps []time.Time
- validatorIDs []types.ValidatorID
+ nodeIDs []types.NodeID
randGen *rand.Rand
hashBlock hashBlockFn
}
-func newValidatorSetStatus(vIDs []types.ValidatorID, hashBlock hashBlockFn) *validatorSetStatus {
- status := make(map[types.ValidatorID]*validatorStatus)
- timestamps := make([]time.Time, 0, len(vIDs))
- proposerChain := make(map[types.ValidatorID]uint32)
- for i, vID := range vIDs {
- status[vID] = &validatorStatus{
+func newNodeSetStatus(nIDs []types.NodeID, hashBlock hashBlockFn) *nodeSetStatus {
+ status := make(map[types.NodeID]*nodeStatus)
+ timestamps := make([]time.Time, 0, len(nIDs))
+ proposerChain := make(map[types.NodeID]uint32)
+ for i, nID := range nIDs {
+ status[nID] = &nodeStatus{
blocks: []*types.Block{},
- lastAckingHeight: make(map[types.ValidatorID]uint64),
+ lastAckingHeight: make(map[types.NodeID]uint64),
}
timestamps = append(timestamps, time.Now().UTC())
- proposerChain[vID] = uint32(i)
+ proposerChain[nID] = uint32(i)
}
- return &validatorSetStatus{
+ return &nodeSetStatus{
status: status,
proposerChain: proposerChain,
timestamps: timestamps,
- validatorIDs: vIDs,
+ nodeIDs: nIDs,
randGen: rand.New(rand.NewSource(time.Now().UnixNano())),
hashBlock: hashBlock,
}
}
-// findIncompleteValidators is a helper to check which validator
+// findIncompleteNodes is a helper to check which node
// doesn't generate enough blocks.
-func (vs *validatorSetStatus) findIncompleteValidators(
- blockCount int) (vIDs []types.ValidatorID) {
+func (vs *nodeSetStatus) findIncompleteNodes(
+ blockCount int) (nIDs []types.NodeID) {
- for vID, status := range vs.status {
+ for nID, status := range vs.status {
if len(status.blocks) < blockCount {
- vIDs = append(vIDs, vID)
+ nIDs = append(nIDs, nID)
}
}
return
}
// prepareAcksForNewBlock collects acks for one block.
-func (vs *validatorSetStatus) prepareAcksForNewBlock(
- proposerID types.ValidatorID, ackingCount int) (
+func (vs *nodeSetStatus) prepareAcksForNewBlock(
+ proposerID types.NodeID, ackingCount int) (
acks common.Hashes, err error) {
acks = common.Hashes{}
@@ -123,22 +126,22 @@ func (vs *validatorSetStatus) prepareAcksForNewBlock(
// The 'Acks' filed of genesis blocks would always be empty.
return
}
- // Pick validatorIDs to be acked.
- ackingVIDs := map[types.ValidatorID]struct{}{
+ // Pick nodeIDs to be acked.
+ ackingNIDs := map[types.NodeID]struct{}{
proposerID: struct{}{}, // Acking parent block is always required.
}
if ackingCount > 0 {
ackingCount-- // We would always include ack to parent block.
}
- for _, i := range vs.randGen.Perm(len(vs.validatorIDs))[:ackingCount] {
- ackingVIDs[vs.validatorIDs[i]] = struct{}{}
+ for _, i := range vs.randGen.Perm(len(vs.nodeIDs))[:ackingCount] {
+ ackingNIDs[vs.nodeIDs[i]] = struct{}{}
}
// Generate acks.
- for vID := range ackingVIDs {
+ for nID := range ackingNIDs {
ack, ok := vs.status[proposerID].getAckedBlockHash(
- vID, vs.status[vID], vs.randGen)
+ nID, vs.status[nID], vs.randGen)
if !ok {
- if vID == proposerID {
+ if nID == proposerID {
err = ErrParentNotAcked
}
continue
@@ -148,9 +151,9 @@ func (vs *validatorSetStatus) prepareAcksForNewBlock(
return
}
-// proposeBlock propose new block and update validator status.
-func (vs *validatorSetStatus) proposeBlock(
- proposerID types.ValidatorID,
+// proposeBlock propose new block and update node status.
+func (vs *nodeSetStatus) proposeBlock(
+ proposerID types.NodeID,
acks common.Hashes) (*types.Block, error) {
status := vs.status[proposerID]
@@ -171,8 +174,8 @@ func (vs *validatorSetStatus) proposeBlock(
Acks: common.NewSortedHashes(acks),
Timestamp: vs.timestamps[chainID],
}
- for i, vID := range vs.validatorIDs {
- if vID == proposerID {
+ for i, nID := range vs.nodeIDs {
+ if nID == proposerID {
newBlock.Position.ChainID = uint32(i)
}
}
@@ -188,13 +191,13 @@ func (vs *validatorSetStatus) proposeBlock(
// normalAckingCountGenerator would randomly pick acking count
// by a normal distribution.
func normalAckingCountGenerator(
- validatorCount int, mean, deviation float64) func() int {
+ nodeCount int, mean, deviation float64) func() int {
return func() int {
var expected float64
for {
expected = rand.NormFloat64()*deviation + mean
- if expected >= 0 && expected <= float64(validatorCount) {
+ if expected >= 0 && expected <= float64(nodeCount) {
break
}
}
@@ -208,32 +211,32 @@ func MaxAckingCountGenerator(count int) func() int {
return func() int { return count }
}
-// generateValidatorPicker is a function generator, which would generate
-// a function to randomly pick one validator ID from a slice of validator ID.
-func generateValidatorPicker() func([]types.ValidatorID) types.ValidatorID {
+// generateNodePicker is a function generator, which would generate
+// a function to randomly pick one node ID from a slice of node ID.
+func generateNodePicker() func([]types.NodeID) types.NodeID {
privateRand := rand.New(rand.NewSource(time.Now().UnixNano()))
- return func(vIDs []types.ValidatorID) types.ValidatorID {
- return vIDs[privateRand.Intn(len(vIDs))]
+ return func(nIDs []types.NodeID) types.NodeID {
+ return nIDs[privateRand.Intn(len(nIDs))]
}
}
// BlocksGenerator could generate blocks forming valid DAGs.
type BlocksGenerator struct {
- validatorPicker func([]types.ValidatorID) types.ValidatorID
- hashBlock hashBlockFn
+ nodePicker func([]types.NodeID) types.NodeID
+ hashBlock hashBlockFn
}
// NewBlocksGenerator constructs BlockGenerator.
-func NewBlocksGenerator(validatorPicker func(
- []types.ValidatorID) types.ValidatorID,
+func NewBlocksGenerator(nodePicker func(
+ []types.NodeID) types.NodeID,
hashBlock hashBlockFn) *BlocksGenerator {
- if validatorPicker == nil {
- validatorPicker = generateValidatorPicker()
+ if nodePicker == nil {
+ nodePicker = generateNodePicker()
}
return &BlocksGenerator{
- validatorPicker: validatorPicker,
- hashBlock: hashBlock,
+ nodePicker: nodePicker,
+ hashBlock: hashBlock,
}
}
@@ -244,45 +247,45 @@ func NewBlocksGenerator(validatorPicker func(
// has maximum 2 acks.
// func () int { return 2 }
// The default ackingCountGenerator would randomly pick a number based on
-// the validatorCount you provided with a normal distribution.
+// the nodeCount you provided with a normal distribution.
func (gen *BlocksGenerator) Generate(
- validatorCount int,
+ nodeCount int,
blockCount int,
ackingCountGenerator func() int,
writer blockdb.Writer) (
- validators types.ValidatorIDs, err error) {
+ nodes types.NodeIDs, err error) {
if ackingCountGenerator == nil {
ackingCountGenerator = normalAckingCountGenerator(
- validatorCount,
- float64(validatorCount/2),
- float64(validatorCount/4+1))
+ nodeCount,
+ float64(nodeCount/2),
+ float64(nodeCount/4+1))
}
- validators = types.ValidatorIDs{}
- for i := 0; i < validatorCount; i++ {
- validators = append(
- validators, types.ValidatorID{Hash: common.NewRandomHash()})
+ nodes = types.NodeIDs{}
+ for i := 0; i < nodeCount; i++ {
+ nodes = append(
+ nodes, types.NodeID{Hash: common.NewRandomHash()})
}
- status := newValidatorSetStatus(validators, gen.hashBlock)
+ status := newNodeSetStatus(nodes, gen.hashBlock)
// We would record the smallest height of block that could be acked
- // from each validator's point-of-view.
- toAck := make(map[types.ValidatorID]map[types.ValidatorID]uint64)
- for _, vID := range validators {
- toAck[vID] = make(map[types.ValidatorID]uint64)
+ // from each node's point-of-view.
+ toAck := make(map[types.NodeID]map[types.NodeID]uint64)
+ for _, nID := range nodes {
+ toAck[nID] = make(map[types.NodeID]uint64)
}
for {
- // Find validators that doesn't propose enough blocks and
+ // Find nodes that doesn't propose enough blocks and
// pick one from them randomly.
- notYet := status.findIncompleteValidators(blockCount)
+ notYet := status.findIncompleteNodes(blockCount)
if len(notYet) == 0 {
break
}
// Propose a new block.
var (
- proposerID = gen.validatorPicker(notYet)
+ proposerID = gen.nodePicker(notYet)
acks common.Hashes
)
acks, err = status.prepareAcksForNewBlock(
diff --git a/core/test/blocks-generator_test.go b/core/test/blocks-generator_test.go
index e607796..f8aa325 100644
--- a/core/test/blocks-generator_test.go
+++ b/core/test/blocks-generator_test.go
@@ -33,21 +33,21 @@ type BlocksGeneratorTestCase struct {
func (s *BlocksGeneratorTestCase) TestGenerate() {
// This test case is to make sure the generated blocks are legimate.
- validatorCount := 19
+ nodeCount := 19
blockCount := 50
gen := NewBlocksGenerator(nil, stableRandomHash)
db, err := blockdb.NewMemBackedBlockDB()
s.Require().Nil(err)
- validators, err := gen.Generate(
- validatorCount, blockCount, nil, db)
+ nodes, err := gen.Generate(
+ nodeCount, blockCount, nil, db)
s.Require().Nil(err)
- s.Require().Len(validators, validatorCount)
+ s.Require().Len(nodes, nodeCount)
// Load all blocks in that database for further checking.
iter, err := db.GetAll()
s.Require().Nil(err)
- blocksByValidator := make(map[types.ValidatorID][]*types.Block)
+ blocksByNode := make(map[types.NodeID][]*types.Block)
blocksByHash := make(map[common.Hash]*types.Block)
for {
block, err := iter.Next()
@@ -56,9 +56,9 @@ func (s *BlocksGeneratorTestCase) TestGenerate() {
}
s.Nil(err)
- blocksByValidator[block.ProposerID] =
- append(blocksByValidator[block.ProposerID], &block)
- sort.Sort(types.ByHeight(blocksByValidator[block.ProposerID]))
+ blocksByNode[block.ProposerID] =
+ append(blocksByNode[block.ProposerID], &block)
+ sort.Sort(types.ByHeight(blocksByNode[block.ProposerID]))
blocksByHash[block.Hash] = &block
}
@@ -67,8 +67,8 @@ func (s *BlocksGeneratorTestCase) TestGenerate() {
// compared to its parent block.
// - Parent Ack: always ack its parent block.
// - No Acks in genesis bloc
- for _, blocks := range blocksByValidator {
- lastAckingHeights := map[types.ValidatorID]uint64{}
+ for _, blocks := range blocksByNode {
+ lastAckingHeights := map[types.NodeID]uint64{}
s.Require().NotEmpty(blocks)
// Check genesis block.
@@ -106,19 +106,19 @@ func (s *BlocksGeneratorTestCase) TestGenerate() {
func (s *BlocksGeneratorTestCase) TestGenerateWithMaxAckCount() {
var (
- validatorCount = 13
- blockCount = 50
- gen = NewBlocksGenerator(nil, stableRandomHash)
- req = s.Require()
+ nodeCount = 13
+ blockCount = 50
+ gen = NewBlocksGenerator(nil, stableRandomHash)
+ req = s.Require()
)
// Generate with 0 acks.
db, err := blockdb.NewMemBackedBlockDB()
req.Nil(err)
- validators, err := gen.Generate(
- validatorCount, blockCount, MaxAckingCountGenerator(0), db)
+ nodes, err := gen.Generate(
+ nodeCount, blockCount, MaxAckingCountGenerator(0), db)
req.Nil(err)
- req.Len(validators, validatorCount)
+ req.Len(nodes, nodeCount)
// Load blocks to check their acking count.
iter, err := db.GetAll()
req.Nil(err)
@@ -137,11 +137,11 @@ func (s *BlocksGeneratorTestCase) TestGenerateWithMaxAckCount() {
// Generate with acks as many as possible.
db, err = blockdb.NewMemBackedBlockDB()
req.Nil(err)
- validators, err = gen.Generate(
- validatorCount, blockCount, MaxAckingCountGenerator(
- validatorCount), db)
+ nodes, err = gen.Generate(
+ nodeCount, blockCount, MaxAckingCountGenerator(
+ nodeCount), db)
req.Nil(err)
- req.Len(validators, validatorCount)
+ req.Len(nodes, nodeCount)
// Load blocks to verify the average acking count.
totalAckingCount := 0
totalBlockCount := 0
@@ -160,7 +160,7 @@ func (s *BlocksGeneratorTestCase) TestGenerateWithMaxAckCount() {
totalBlockCount++
}
req.NotZero(totalBlockCount)
- req.True((totalAckingCount / totalBlockCount) >= (validatorCount / 2))
+ req.True((totalAckingCount / totalBlockCount) >= (nodeCount / 2))
}
func TestBlocksGenerator(t *testing.T) {
diff --git a/core/test/fake-transport.go b/core/test/fake-transport.go
index 2615bd4..2f1686e 100644
--- a/core/test/fake-transport.go
+++ b/core/test/fake-transport.go
@@ -28,10 +28,10 @@ import (
// by using golang channel.
type FakeTransport struct {
peerType TransportPeerType
- vID types.ValidatorID
+ nID types.NodeID
recvChannel chan *TransportEnvelope
serverChannel chan<- *TransportEnvelope
- peers map[types.ValidatorID]chan<- *TransportEnvelope
+ peers map[types.NodeID]chan<- *TransportEnvelope
latency LatencyModel
}
@@ -45,19 +45,19 @@ func NewFakeTransportServer() TransportServer {
// NewFakeTransportClient constructs FakeTransport instance for peer.
func NewFakeTransportClient(
- vID types.ValidatorID, latency LatencyModel) TransportClient {
+ nID types.NodeID, latency LatencyModel) TransportClient {
return &FakeTransport{
peerType: TransportPeer,
recvChannel: make(chan *TransportEnvelope, 1000),
- vID: vID,
+ nID: nID,
latency: latency,
}
}
// Send implements Transport.Send method.
func (t *FakeTransport) Send(
- endpoint types.ValidatorID, msg interface{}) (err error) {
+ endpoint types.NodeID, msg interface{}) (err error) {
ch, exists := t.peers[endpoint]
if !exists {
@@ -70,7 +70,7 @@ func (t *FakeTransport) Send(
}
ch <- &TransportEnvelope{
PeerType: t.peerType,
- From: t.vID,
+ From: t.nID,
Msg: msg,
}
}(ch)
@@ -82,7 +82,7 @@ func (t *FakeTransport) Report(msg interface{}) (err error) {
go func() {
t.serverChannel <- &TransportEnvelope{
PeerType: TransportPeer,
- From: t.vID,
+ From: t.nID,
Msg: msg,
}
}()
@@ -92,7 +92,7 @@ func (t *FakeTransport) Report(msg interface{}) (err error) {
// Broadcast implements Transport.Broadcast method.
func (t *FakeTransport) Broadcast(msg interface{}) (err error) {
for k := range t.peers {
- if k == t.vID {
+ if k == t.nID {
continue
}
t.Send(k, msg)
@@ -107,10 +107,10 @@ func (t *FakeTransport) Close() (err error) {
}
// Peers implements Transport.Peers method.
-func (t *FakeTransport) Peers() (peers map[types.ValidatorID]struct{}) {
- peers = make(map[types.ValidatorID]struct{})
- for vID := range t.peers {
- peers[vID] = struct{}{}
+func (t *FakeTransport) Peers() (peers map[types.NodeID]struct{}) {
+ peers = make(map[types.NodeID]struct{})
+ for nID := range t.peers {
+ peers[nID] = struct{}{}
}
return
}
@@ -135,7 +135,7 @@ func (t *FakeTransport) Join(
continue
}
if t.peers, ok =
- envelope.Msg.(map[types.ValidatorID]chan<- *TransportEnvelope); !ok {
+ envelope.Msg.(map[types.NodeID]chan<- *TransportEnvelope); !ok {
envelopes = append(envelopes, envelope)
continue
@@ -155,7 +155,7 @@ func (t *FakeTransport) Host() (chan *TransportEnvelope, error) {
// WaitForPeers implements TransportServer.WaitForPeers method.
func (t *FakeTransport) WaitForPeers(numPeers int) (err error) {
- t.peers = make(map[types.ValidatorID]chan<- *TransportEnvelope)
+ t.peers = make(map[types.NodeID]chan<- *TransportEnvelope)
for {
envelope := <-t.recvChannel
// Panic here if some peer send other stuffs before
diff --git a/core/test/governance.go b/core/test/governance.go
index c5746cb..a2e6f69 100644
--- a/core/test/governance.go
+++ b/core/test/governance.go
@@ -28,44 +28,44 @@ import (
var (
// ErrPrivateKeyNotExists means caller request private key for an
- // unknown validator ID.
+ // unknown node ID.
ErrPrivateKeyNotExists = fmt.Errorf("private key not exists")
)
// Governance is an implementation of Goverance for testing purpose.
type Governance struct {
lambda time.Duration
- notarySet map[types.ValidatorID]struct{}
- privateKeys map[types.ValidatorID]crypto.PrivateKey
+ notarySet map[types.NodeID]struct{}
+ privateKeys map[types.NodeID]crypto.PrivateKey
DKGComplaint map[uint64][]*types.DKGComplaint
DKGMasterPublicKey map[uint64][]*types.DKGMasterPublicKey
}
// NewGovernance constructs a Governance instance.
-func NewGovernance(validatorCount int, lambda time.Duration) (
+func NewGovernance(nodeCount int, lambda time.Duration) (
g *Governance, err error) {
g = &Governance{
lambda: lambda,
- notarySet: make(map[types.ValidatorID]struct{}),
- privateKeys: make(map[types.ValidatorID]crypto.PrivateKey),
+ notarySet: make(map[types.NodeID]struct{}),
+ privateKeys: make(map[types.NodeID]crypto.PrivateKey),
DKGComplaint: make(map[uint64][]*types.DKGComplaint),
DKGMasterPublicKey: make(map[uint64][]*types.DKGMasterPublicKey),
}
- for i := 0; i < validatorCount; i++ {
+ for i := 0; i < nodeCount; i++ {
prv, err := eth.NewPrivateKey()
if err != nil {
return nil, err
}
- vID := types.NewValidatorID(prv.PublicKey())
- g.notarySet[vID] = struct{}{}
- g.privateKeys[vID] = prv
+ nID := types.NewNodeID(prv.PublicKey())
+ g.notarySet[nID] = struct{}{}
+ g.privateKeys[nID] = prv
}
return
}
// GetNotarySet implements Governance interface to return current
// notary set.
-func (g *Governance) GetNotarySet() map[types.ValidatorID]struct{} {
+func (g *Governance) GetNotarySet() map[types.NodeID]struct{} {
return g.notarySet
}
@@ -81,12 +81,12 @@ func (g *Governance) GetConfiguration(blockHeight uint64) *types.Config {
}
}
-// GetPrivateKey return the private key for that validator, this function
+// GetPrivateKey return the private key for that node, this function
// is a test utility and not a general Governance interface.
func (g *Governance) GetPrivateKey(
- vID types.ValidatorID) (key crypto.PrivateKey, err error) {
+ nID types.NodeID) (key crypto.PrivateKey, err error) {
- key, exists := g.privateKeys[vID]
+ key, exists := g.privateKeys[nID]
if !exists {
err = ErrPrivateKeyNotExists
return
diff --git a/core/test/interface.go b/core/test/interface.go
index 0dc2382..a422ee7 100644
--- a/core/test/interface.go
+++ b/core/test/interface.go
@@ -38,7 +38,7 @@ type Stopper interface {
// moment.
// The Stopper should check state of that handler and return 'true'
// if the execution could be stopped.
- ShouldStop(vID types.ValidatorID) bool
+ ShouldStop(nID types.NodeID) bool
}
// EventHandler defines an interface to handle a Scheduler event.
@@ -62,8 +62,8 @@ type TransportEnvelope struct {
// PeerType defines the type of source peer, could be either "peer" or
// "server".
PeerType TransportPeerType
- // From defines the validatorID of the source peer.
- From types.ValidatorID
+ // From defines the nodeID of the source peer.
+ From types.NodeID
// Msg is the actual payload of this message.
Msg interface{}
}
@@ -92,14 +92,14 @@ type Transport interface {
// Broadcast a message to all peers in network.
Broadcast(msg interface{}) error
// Send one message to a peer.
- Send(endpoint types.ValidatorID, msg interface{}) error
+ Send(endpoint types.NodeID, msg interface{}) error
// Close would cleanup allocated resources.
Close() error
- // Peers return IDs of all connected validators in p2p favor.
+ // Peers return IDs of all connected nodes in p2p favor.
// This method should be accessed after ether 'Join' or 'WaitForPeers'
// returned.
- Peers() map[types.ValidatorID]struct{}
+ Peers() map[types.NodeID]struct{}
}
// Marshaller defines an interface to convert between interface{} and []byte.
diff --git a/core/test/revealer.go b/core/test/revealer.go
index b8eb9b4..b3af4d7 100644
--- a/core/test/revealer.go
+++ b/core/test/revealer.go
@@ -63,12 +63,12 @@ func loadAllBlocks(iter blockdb.BlockIterator) (
// all blocks from blockdb, and randomly pick one block to reveal if
// it still forms a valid DAG in revealed blocks.
type RandomDAGRevealer struct {
- // blocksByValidator group all blocks by validators and sorting
+ // blocksByNode group all blocks by nodes and sorting
// them by height.
- blocksByValidator map[types.ValidatorID][]*types.Block
- // tipIndexes store the height of next block from one validator
+ blocksByNode map[types.NodeID][]*types.Block
+ // tipIndexes store the height of next block from one node
// to check if is candidate.
- tipIndexes map[types.ValidatorID]int
+ tipIndexes map[types.NodeID]int
// candidate are blocks that forms valid DAG with
// current revealed blocks.
candidates []*types.Block
@@ -86,19 +86,19 @@ func NewRandomDAGRevealer(
return
}
- // Rearrange blocks by validators and height.
- blocksByValidator := make(map[types.ValidatorID][]*types.Block)
+ // Rearrange blocks by nodes and height.
+ blocksByNode := make(map[types.NodeID][]*types.Block)
for _, block := range blocks {
- blocksByValidator[block.ProposerID] =
- append(blocksByValidator[block.ProposerID], block)
+ blocksByNode[block.ProposerID] =
+ append(blocksByNode[block.ProposerID], block)
}
// Make sure blocks are sorted by block heights, from lower to higher.
- for vID := range blocksByValidator {
- sort.Sort(types.ByHeight(blocksByValidator[vID]))
+ for nID := range blocksByNode {
+ sort.Sort(types.ByHeight(blocksByNode[nID]))
}
r = &RandomDAGRevealer{
- blocksByValidator: blocksByValidator,
- randGen: rand.New(rand.NewSource(time.Now().UnixNano())),
+ blocksByNode: blocksByNode,
+ randGen: rand.New(rand.NewSource(time.Now().UnixNano())),
}
// Make sure this revealer is ready to use.
r.Reset()
@@ -107,8 +107,8 @@ func NewRandomDAGRevealer(
// pickCandidates is a helper function to pick candidates from current tips.
func (r *RandomDAGRevealer) pickCandidates() {
- for vID, tip := range r.tipIndexes {
- blocks, exists := r.blocksByValidator[vID]
+ for nID, tip := range r.tipIndexes {
+ blocks, exists := r.blocksByNode[nID]
if !exists {
continue
}
@@ -117,7 +117,7 @@ func (r *RandomDAGRevealer) pickCandidates() {
}
block := blocks[tip]
if isAllAckingBlockRevealed(block, r.revealed) {
- r.tipIndexes[vID]++
+ r.tipIndexes[nID]++
r.candidates = append(r.candidates, block)
}
}
@@ -145,9 +145,9 @@ func (r *RandomDAGRevealer) Next() (types.Block, error) {
// Reset implement Revealer.Reset method, which would reset the revealing.
func (r *RandomDAGRevealer) Reset() {
- r.tipIndexes = make(map[types.ValidatorID]int)
- for vID := range r.blocksByValidator {
- r.tipIndexes[vID] = 0
+ r.tipIndexes = make(map[types.NodeID]int)
+ for nID := range r.blocksByNode {
+ r.tipIndexes[nID] = 0
}
r.revealed = make(map[common.Hash]struct{})
r.candidates = []*types.Block{}
diff --git a/core/test/revealer_test.go b/core/test/revealer_test.go
index 16d3b18..0c20520 100644
--- a/core/test/revealer_test.go
+++ b/core/test/revealer_test.go
@@ -35,9 +35,9 @@ type RevealerTestSuite struct {
func (s *RevealerTestSuite) SetupSuite() {
var (
- err error
- validatorCount = 19
- blockCount = 50
+ err error
+ nodeCount = 19
+ blockCount = 50
)
// Setup block database.
s.db, err = blockdb.NewMemBackedBlockDB()
@@ -45,10 +45,10 @@ func (s *RevealerTestSuite) SetupSuite() {
// Randomly generate blocks.
gen := NewBlocksGenerator(nil, stableRandomHash)
- validators, err := gen.Generate(
- validatorCount, blockCount, nil, s.db)
+ nodes, err := gen.Generate(
+ nodeCount, blockCount, nil, s.db)
s.Require().Nil(err)
- s.Require().Len(validators, validatorCount)
+ s.Require().Len(nodes, nodeCount)
// Cache the count of total generated block.
iter, err := s.db.GetAll()
diff --git a/core/test/scheduler-event.go b/core/test/scheduler-event.go
index 85968c5..2863b54 100644
--- a/core/test/scheduler-event.go
+++ b/core/test/scheduler-event.go
@@ -27,8 +27,8 @@ import (
type Event struct {
// HistoryIndex is the index of this event in history.
HistoryIndex int
- // ValidatorID is the ID of handler that this event deginated to.
- ValidatorID types.ValidatorID
+ // NodeID is the ID of handler that this event deginated to.
+ NodeID types.NodeID
// Time is the expected execution time of this event.
Time time.Time
// ExecError record the error when handling this event.
@@ -67,12 +67,12 @@ func (eq *eventQueue) Pop() interface{} {
// NewEvent is the constructor for Event.
func NewEvent(
- vID types.ValidatorID, when time.Time, payload interface{}) *Event {
+ nID types.NodeID, when time.Time, payload interface{}) *Event {
return &Event{
HistoryIndex: -1,
ParentHistoryIndex: -1,
- ValidatorID: vID,
+ NodeID: nID,
Time: when,
Payload: payload,
}
diff --git a/core/test/scheduler.go b/core/test/scheduler.go
index 6a3a40a..7c5bbde 100644
--- a/core/test/scheduler.go
+++ b/core/test/scheduler.go
@@ -48,7 +48,7 @@ type Scheduler struct {
history []*Event
historyLock sync.RWMutex
isStarted bool
- handlers map[types.ValidatorID]*schedulerHandlerRecord
+ handlers map[types.NodeID]*schedulerHandlerRecord
handlersLock sync.RWMutex
eventNotification chan struct{}
ctx context.Context
@@ -62,7 +62,7 @@ func NewScheduler(stopper Stopper) *Scheduler {
return &Scheduler{
events: eventQueue{},
history: []*Event{},
- handlers: make(map[types.ValidatorID]*schedulerHandlerRecord),
+ handlers: make(map[types.NodeID]*schedulerHandlerRecord),
eventNotification: make(chan struct{}, 100000),
ctx: ctx,
cancelFunc: cancel,
@@ -98,15 +98,15 @@ func (sch *Scheduler) Seed(e *Event) error {
}
// RegisterEventHandler register an event handler by providing ID of
-// corresponding validator.
+// corresponding node.
func (sch *Scheduler) RegisterEventHandler(
- vID types.ValidatorID,
+ nID types.NodeID,
handler EventHandler) {
sch.handlersLock.Lock()
defer sch.handlersLock.Unlock()
- sch.handlers[vID] = &schedulerHandlerRecord{handler: handler}
+ sch.handlers[nID] = &schedulerHandlerRecord{handler: handler}
}
// nextTick would pick the oldest event from eventQueue.
@@ -144,12 +144,12 @@ func (sch *Scheduler) workerRoutine(wg *sync.WaitGroup) {
handleEvent := func(e *Event) {
// Find correspond handler record.
- hRec := func(vID types.ValidatorID) *schedulerHandlerRecord {
+ hRec := func(nID types.NodeID) *schedulerHandlerRecord {
sch.handlersLock.RLock()
defer sch.handlersLock.RUnlock()
- return sch.handlers[vID]
- }(e.ValidatorID)
+ return sch.handlers[nID]
+ }(e.NodeID)
newEvents := func() []*Event {
// This lock makes sure there would be no concurrent access
@@ -161,8 +161,8 @@ func (sch *Scheduler) workerRoutine(wg *sync.WaitGroup) {
beforeExecution := time.Now().UTC()
newEvents := hRec.handler.Handle(e)
e.ExecInterval = time.Now().UTC().Sub(beforeExecution)
- // It's safe to check status of that validator under 'hRec.lock'.
- if sch.stopper.ShouldStop(e.ValidatorID) {
+ // It's safe to check status of that node under 'hRec.lock'.
+ if sch.stopper.ShouldStop(e.NodeID) {
sch.cancelFunc()
}
return newEvents
diff --git a/core/test/scheduler_test.go b/core/test/scheduler_test.go
index 5aef36e..1e6d52f 100644
--- a/core/test/scheduler_test.go
+++ b/core/test/scheduler_test.go
@@ -33,16 +33,16 @@ type SchedulerTestSuite struct {
type simpleStopper struct {
lock sync.Mutex
- touched map[types.ValidatorID]int
+ touched map[types.NodeID]int
touchedCount int
}
func newSimpleStopper(
- validators []types.ValidatorID, touchedCount int) *simpleStopper {
+ nodes []types.NodeID, touchedCount int) *simpleStopper {
- touched := make(map[types.ValidatorID]int)
- for _, vID := range validators {
- touched[vID] = 0
+ touched := make(map[types.NodeID]int)
+ for _, nID := range nodes {
+ touched[nID] = 0
}
return &simpleStopper{
touched: touched,
@@ -50,11 +50,11 @@ func newSimpleStopper(
}
}
-func (stopper *simpleStopper) ShouldStop(vID types.ValidatorID) bool {
+func (stopper *simpleStopper) ShouldStop(nID types.NodeID) bool {
stopper.lock.Lock()
defer stopper.lock.Unlock()
- stopper.touched[vID] = stopper.touched[vID] + 1
+ stopper.touched[nID] = stopper.touched[nID] + 1
for _, count := range stopper.touched {
if count < stopper.touchedCount {
return false
@@ -65,26 +65,26 @@ func (stopper *simpleStopper) ShouldStop(vID types.ValidatorID) bool {
type simpleHandler struct {
count int
- vID types.ValidatorID
+ nID types.NodeID
}
func (handler *simpleHandler) Handle(e *Event) (events []*Event) {
- if e.ValidatorID == handler.vID {
+ if e.NodeID == handler.nID {
handler.count++
}
return
}
type fixedLatencyHandler struct {
- vID types.ValidatorID
+ nID types.NodeID
}
func (handler *fixedLatencyHandler) Handle(e *Event) (events []*Event) {
// Simulate execution time.
time.Sleep(500 * time.Millisecond)
return []*Event{&Event{
- ValidatorID: handler.vID,
- Time: e.Time.Add(800 * time.Millisecond),
+ NodeID: handler.nID,
+ Time: e.Time.Add(800 * time.Millisecond),
}}
}
@@ -114,22 +114,22 @@ func (s *SchedulerTestSuite) TestEventSequence() {
func (s *SchedulerTestSuite) TestBasicRound() {
// This test case makes sure these facts:
- // - event is dispatched by validatorID attached to each handler.
+ // - event is dispatched by NodeID attached to each handler.
// - stopper can stop the execution when condition is met.
var (
- req = s.Require()
- validators = GenerateRandomValidatorIDs(3)
- stopper = newSimpleStopper(validators, 2)
- sch = NewScheduler(stopper)
- handlers = make(map[types.ValidatorID]*simpleHandler)
+ req = s.Require()
+ nodes = GenerateRandomNodeIDs(3)
+ stopper = newSimpleStopper(nodes, 2)
+ sch = NewScheduler(stopper)
+ handlers = make(map[types.NodeID]*simpleHandler)
)
- for _, vID := range validators {
- handler := &simpleHandler{vID: vID}
- handlers[vID] = handler
- sch.RegisterEventHandler(vID, handler)
- req.Nil(sch.Seed(&Event{ValidatorID: vID}))
- req.Nil(sch.Seed(&Event{ValidatorID: vID}))
+ for _, nID := range nodes {
+ handler := &simpleHandler{nID: nID}
+ handlers[nID] = handler
+ sch.RegisterEventHandler(nID, handler)
+ req.Nil(sch.Seed(&Event{NodeID: nID}))
+ req.Nil(sch.Seed(&Event{NodeID: nID}))
}
sch.Run(10)
// Verify result.
@@ -143,16 +143,16 @@ func (s *SchedulerTestSuite) TestChildEvent() {
// assigned correctly.
var (
req = s.Require()
- vID = types.ValidatorID{Hash: common.NewRandomHash()}
- stopper = newSimpleStopper(types.ValidatorIDs{vID}, 3)
- handler = &fixedLatencyHandler{vID: vID}
+ nID = types.NodeID{Hash: common.NewRandomHash()}
+ stopper = newSimpleStopper(types.NodeIDs{nID}, 3)
+ handler = &fixedLatencyHandler{nID: nID}
sch = NewScheduler(stopper)
)
- sch.RegisterEventHandler(vID, handler)
+ sch.RegisterEventHandler(nID, handler)
req.Nil(sch.Seed(&Event{
- ValidatorID: vID,
- Time: time.Now().UTC(),
+ NodeID: nID,
+ Time: time.Now().UTC(),
}))
sch.Run(1)
// Verify result.
diff --git a/core/test/stopper.go b/core/test/stopper.go
index 7c75958..9fe5592 100644
--- a/core/test/stopper.go
+++ b/core/test/stopper.go
@@ -24,13 +24,13 @@ import (
"github.com/dexon-foundation/dexon-consensus-core/core/types"
)
-// StopByConfirmedBlocks would make sure each validators confirms
+// StopByConfirmedBlocks would make sure each nodes confirms
// at least X blocks proposed by itself.
type StopByConfirmedBlocks struct {
- apps map[types.ValidatorID]*App
- dbs map[types.ValidatorID]blockdb.BlockDatabase
- lastCheckDelivered map[types.ValidatorID]int
- confirmedBlocks map[types.ValidatorID]int
+ apps map[types.NodeID]*App
+ dbs map[types.NodeID]blockdb.BlockDatabase
+ lastCheckDelivered map[types.NodeID]int
+ confirmedBlocks map[types.NodeID]int
blockCount int
lock sync.Mutex
}
@@ -38,45 +38,45 @@ type StopByConfirmedBlocks struct {
// NewStopByConfirmedBlocks construct an StopByConfirmedBlocks instance.
func NewStopByConfirmedBlocks(
blockCount int,
- apps map[types.ValidatorID]*App,
- dbs map[types.ValidatorID]blockdb.BlockDatabase) *StopByConfirmedBlocks {
+ apps map[types.NodeID]*App,
+ dbs map[types.NodeID]blockdb.BlockDatabase) *StopByConfirmedBlocks {
- confirmedBlocks := make(map[types.ValidatorID]int)
- for vID := range apps {
- confirmedBlocks[vID] = 0
+ confirmedBlocks := make(map[types.NodeID]int)
+ for nID := range apps {
+ confirmedBlocks[nID] = 0
}
return &StopByConfirmedBlocks{
apps: apps,
dbs: dbs,
- lastCheckDelivered: make(map[types.ValidatorID]int),
+ lastCheckDelivered: make(map[types.NodeID]int),
confirmedBlocks: confirmedBlocks,
blockCount: blockCount,
}
}
// ShouldStop implements Stopper interface.
-func (s *StopByConfirmedBlocks) ShouldStop(vID types.ValidatorID) bool {
+func (s *StopByConfirmedBlocks) ShouldStop(nID types.NodeID) bool {
s.lock.Lock()
defer s.lock.Unlock()
- // Accumulate confirmed blocks proposed by this validator in this round.
- lastChecked := s.lastCheckDelivered[vID]
- currentConfirmedBlocks := s.confirmedBlocks[vID]
- db := s.dbs[vID]
- s.apps[vID].Check(func(app *App) {
+ // Accumulate confirmed blocks proposed by this node in this round.
+ lastChecked := s.lastCheckDelivered[nID]
+ currentConfirmedBlocks := s.confirmedBlocks[nID]
+ db := s.dbs[nID]
+ s.apps[nID].Check(func(app *App) {
for _, h := range app.DeliverSequence[lastChecked:] {
b, err := db.Get(h)
if err != nil {
panic(err)
}
- if b.ProposerID == vID {
+ if b.ProposerID == nID {
currentConfirmedBlocks++
}
}
- s.lastCheckDelivered[vID] = len(app.DeliverSequence)
+ s.lastCheckDelivered[nID] = len(app.DeliverSequence)
})
- s.confirmedBlocks[vID] = currentConfirmedBlocks
- // Check if all validators confirmed at least 'blockCount' blocks.
+ s.confirmedBlocks[nID] = currentConfirmedBlocks
+ // Check if all nodes confirmed at least 'blockCount' blocks.
for _, v := range s.confirmedBlocks {
if v < s.blockCount {
return false
diff --git a/core/test/stopper_test.go b/core/test/stopper_test.go
index 9a0e430..262e178 100644
--- a/core/test/stopper_test.go
+++ b/core/test/stopper_test.go
@@ -33,20 +33,20 @@ type StopperTestSuite struct {
func (s *StopperTestSuite) TestStopByConfirmedBlocks() {
// This test case makes sure this stopper would stop when
- // all validators confirmed at least 'x' count of blocks produced
+ // all nodes confirmed at least 'x' count of blocks produced
// by themselves.
var (
req = s.Require()
)
- apps := make(map[types.ValidatorID]*App)
- dbs := make(map[types.ValidatorID]blockdb.BlockDatabase)
- validators := GenerateRandomValidatorIDs(2)
+ apps := make(map[types.NodeID]*App)
+ dbs := make(map[types.NodeID]blockdb.BlockDatabase)
+ nodes := GenerateRandomNodeIDs(2)
db, err := blockdb.NewMemBackedBlockDB()
req.Nil(err)
- for _, vID := range validators {
- apps[vID] = NewApp()
- dbs[vID] = db
+ for _, nID := range nodes {
+ apps[nID] = NewApp()
+ dbs[nID] = db
}
deliver := func(blocks []*types.Block) {
hashes := common.Hashes{}
@@ -54,8 +54,8 @@ func (s *StopperTestSuite) TestStopByConfirmedBlocks() {
hashes = append(hashes, b.Hash)
req.Nil(db.Put(*b))
}
- for _, vID := range validators {
- app := apps[vID]
+ for _, nID := range nodes {
+ app := apps[nID]
for _, h := range hashes {
app.StronglyAcked(h)
}
@@ -67,35 +67,35 @@ func (s *StopperTestSuite) TestStopByConfirmedBlocks() {
}
stopper := NewStopByConfirmedBlocks(2, apps, dbs)
b00 := &types.Block{
- ProposerID: validators[0],
+ ProposerID: nodes[0],
Hash: common.NewRandomHash(),
}
deliver([]*types.Block{b00})
b10 := &types.Block{
- ProposerID: validators[1],
+ ProposerID: nodes[1],
Hash: common.NewRandomHash(),
}
b11 := &types.Block{
- ProposerID: validators[1],
+ ProposerID: nodes[1],
ParentHash: b10.Hash,
Hash: common.NewRandomHash(),
}
deliver([]*types.Block{b10, b11})
- req.False(stopper.ShouldStop(validators[1]))
+ req.False(stopper.ShouldStop(nodes[1]))
b12 := &types.Block{
- ProposerID: validators[1],
+ ProposerID: nodes[1],
ParentHash: b11.Hash,
Hash: common.NewRandomHash(),
}
deliver([]*types.Block{b12})
- req.False(stopper.ShouldStop(validators[1]))
+ req.False(stopper.ShouldStop(nodes[1]))
b01 := &types.Block{
- ProposerID: validators[0],
+ ProposerID: nodes[0],
ParentHash: b00.Hash,
Hash: common.NewRandomHash(),
}
deliver([]*types.Block{b01})
- req.True(stopper.ShouldStop(validators[0]))
+ req.True(stopper.ShouldStop(nodes[0]))
}
func TestStopper(t *testing.T) {
diff --git a/core/test/tcp-transport.go b/core/test/tcp-transport.go
index 2afea14..8bbaf9c 100644
--- a/core/test/tcp-transport.go
+++ b/core/test/tcp-transport.go
@@ -37,18 +37,18 @@ import (
// tcpMessage is the general message between peers and server.
type tcpMessage struct {
- ValidatorID types.ValidatorID `json:"vid"`
- Type string `json:"type"`
- Info string `json:"conn"`
+ NodeID types.NodeID `json:"nid"`
+ Type string `json:"type"`
+ Info string `json:"conn"`
}
// TCPTransport implements Transport interface via TCP connection.
type TCPTransport struct {
peerType TransportPeerType
- vID types.ValidatorID
+ nID types.NodeID
localPort int
- peersInfo map[types.ValidatorID]string
- peers map[types.ValidatorID]chan<- []byte
+ peersInfo map[types.NodeID]string
+ peers map[types.NodeID]chan<- []byte
peersLock sync.RWMutex
recvChannel chan *TransportEnvelope
ctx context.Context
@@ -60,7 +60,7 @@ type TCPTransport struct {
// NewTCPTransport constructs an TCPTransport instance.
func NewTCPTransport(
peerType TransportPeerType,
- vID types.ValidatorID,
+ nID types.NodeID,
latency LatencyModel,
marshaller Marshaller,
localPort int) *TCPTransport {
@@ -68,9 +68,9 @@ func NewTCPTransport(
ctx, cancel := context.WithCancel(context.Background())
return &TCPTransport{
peerType: peerType,
- vID: vID,
- peersInfo: make(map[types.ValidatorID]string),
- peers: make(map[types.ValidatorID]chan<- []byte),
+ nID: nID,
+ peersInfo: make(map[types.NodeID]string),
+ peers: make(map[types.NodeID]chan<- []byte),
recvChannel: make(chan *TransportEnvelope, 1000),
ctx: ctx,
cancel: cancel,
@@ -82,7 +82,7 @@ func NewTCPTransport(
// Send implements Transport.Send method.
func (t *TCPTransport) Send(
- endpoint types.ValidatorID, msg interface{}) (err error) {
+ endpoint types.NodeID, msg interface{}) (err error) {
payload, err := t.marshalMessage(msg)
if err != nil {
@@ -110,8 +110,8 @@ func (t *TCPTransport) Broadcast(msg interface{}) (err error) {
t.peersLock.RLock()
defer t.peersLock.RUnlock()
- for vID, ch := range t.peers {
- if vID == t.vID {
+ for nID, ch := range t.peers {
+ if nID == t.nID {
continue
}
go func(ch chan<- []byte) {
@@ -131,7 +131,7 @@ func (t *TCPTransport) Close() (err error) {
// Reset peers.
t.peersLock.Lock()
defer t.peersLock.Unlock()
- t.peers = make(map[types.ValidatorID]chan<- []byte)
+ t.peers = make(map[types.NodeID]chan<- []byte)
// Tell our user that this channel is closed.
close(t.recvChannel)
t.recvChannel = nil
@@ -139,10 +139,10 @@ func (t *TCPTransport) Close() (err error) {
}
// Peers implements Transport.Peers method.
-func (t *TCPTransport) Peers() (peers map[types.ValidatorID]struct{}) {
- peers = make(map[types.ValidatorID]struct{})
- for vID := range t.peersInfo {
- peers[vID] = struct{}{}
+func (t *TCPTransport) Peers() (peers map[types.NodeID]struct{}) {
+ peers = make(map[types.NodeID]struct{})
+ for nID := range t.peersInfo {
+ peers[nID] = struct{}{}
}
return
}
@@ -152,16 +152,16 @@ func (t *TCPTransport) marshalMessage(
msgCarrier := struct {
PeerType TransportPeerType `json:"peer_type"`
- From types.ValidatorID `json:"from"`
+ From types.NodeID `json:"from"`
Type string `json:"type"`
Payload interface{} `json:"payload"`
}{
PeerType: t.peerType,
- From: t.vID,
+ From: t.nID,
Payload: msg,
}
switch msg.(type) {
- case map[types.ValidatorID]string:
+ case map[types.NodeID]string:
msgCarrier.Type = "peerlist"
case *tcpMessage:
msgCarrier.Type = "trans-msg"
@@ -188,13 +188,13 @@ func (t *TCPTransport) marshalMessage(
func (t *TCPTransport) unmarshalMessage(
payload []byte) (
peerType TransportPeerType,
- from types.ValidatorID,
+ from types.NodeID,
msg interface{},
err error) {
msgCarrier := struct {
PeerType TransportPeerType `json:"peer_type"`
- From types.ValidatorID `json:"from"`
+ From types.NodeID `json:"from"`
Type string `json:"type"`
Payload json.RawMessage `json:"payload"`
}{}
@@ -205,7 +205,7 @@ func (t *TCPTransport) unmarshalMessage(
from = msgCarrier.From
switch msgCarrier.Type {
case "peerlist":
- var peers map[types.ValidatorID]string
+ var peers map[types.NodeID]string
if err = json.Unmarshal(msgCarrier.Payload, &peers); err != nil {
return
}
@@ -376,12 +376,12 @@ func (t *TCPTransport) listenerRoutine(listener *net.TCPListener) {
// we only utilize the write part for simplicity.
func (t *TCPTransport) buildConnectionsToPeers() (err error) {
var wg sync.WaitGroup
- for vID, addr := range t.peersInfo {
- if vID == t.vID {
+ for nID, addr := range t.peersInfo {
+ if nID == t.nID {
continue
}
wg.Add(1)
- go func(vID types.ValidatorID, addr string) {
+ go func(nID types.NodeID, addr string) {
defer wg.Done()
conn, localErr := net.Dial("tcp", addr)
@@ -394,8 +394,8 @@ func (t *TCPTransport) buildConnectionsToPeers() (err error) {
t.peersLock.Lock()
defer t.peersLock.Unlock()
- t.peers[vID] = t.connWriter(conn)
- }(vID, addr)
+ t.peers[nID] = t.connWriter(conn)
+ }(nID, addr)
}
wg.Wait()
return
@@ -410,13 +410,13 @@ type TCPTransportClient struct {
// NewTCPTransportClient constructs a TCPTransportClient instance.
func NewTCPTransportClient(
- vID types.ValidatorID,
+ nID types.NodeID,
latency LatencyModel,
marshaller Marshaller,
local bool) *TCPTransportClient {
return &TCPTransportClient{
- TCPTransport: *NewTCPTransport(TransportPeer, vID, latency, marshaller, 8080),
+ TCPTransport: *NewTCPTransport(TransportPeer, nID, latency, marshaller, 8080),
local: local,
}
}
@@ -492,15 +492,15 @@ func (t *TCPTransportClient) Join(
conn = net.JoinHostPort(ip, strconv.Itoa(t.localPort))
}
if err = t.Report(&tcpMessage{
- Type: "conn",
- ValidatorID: t.vID,
- Info: conn,
+ Type: "conn",
+ NodeID: t.nID,
+ Info: conn,
}); err != nil {
return
}
// Wait for peers list sent by server.
e := <-t.recvChannel
- if t.peersInfo, ok = e.Msg.(map[types.ValidatorID]string); !ok {
+ if t.peersInfo, ok = e.Msg.(map[types.NodeID]string); !ok {
panic(fmt.Errorf("expect peer list, not %v", e))
}
// Setup connections to other peers.
@@ -509,8 +509,8 @@ func (t *TCPTransportClient) Join(
}
// Report to server that the connections to other peers are ready.
if err = t.Report(&tcpMessage{
- Type: "conn-ready",
- ValidatorID: t.vID,
+ Type: "conn-ready",
+ NodeID: t.nID,
}); err != nil {
return
}
@@ -547,11 +547,11 @@ func NewTCPTransportServer(
serverPort int) *TCPTransportServer {
return &TCPTransportServer{
- // NOTE: the assumption here is the validator ID of peers
+ // NOTE: the assumption here is the node ID of peers
// won't be zero.
TCPTransport: *NewTCPTransport(
TransportPeerServer,
- types.ValidatorID{},
+ types.NodeID{},
nil,
marshaller,
serverPort),
@@ -586,7 +586,7 @@ func (t *TCPTransportServer) WaitForPeers(numPeers int) (err error) {
if msg.Type != "conn" {
panic(fmt.Errorf("expect connection report, not %v", e))
}
- t.peersInfo[msg.ValidatorID] = msg.Info
+ t.peersInfo[msg.NodeID] = msg.Info
// Check if we already collect enought peers.
if len(t.peersInfo) == numPeers {
break
@@ -600,7 +600,7 @@ func (t *TCPTransportServer) WaitForPeers(numPeers int) (err error) {
return
}
// Wait for peers to send 'ready' report.
- readies := make(map[types.ValidatorID]struct{})
+ readies := make(map[types.NodeID]struct{})
for {
e := <-t.recvChannel
msg, ok := e.Msg.(*tcpMessage)
@@ -610,10 +610,10 @@ func (t *TCPTransportServer) WaitForPeers(numPeers int) (err error) {
if msg.Type != "conn-ready" {
panic(fmt.Errorf("expect connection ready, not %v", e))
}
- if _, reported := readies[msg.ValidatorID]; reported {
+ if _, reported := readies[msg.NodeID]; reported {
panic(fmt.Errorf("already report conn-ready message: %v", e))
}
- readies[msg.ValidatorID] = struct{}{}
+ readies[msg.NodeID] = struct{}{}
if len(readies) == numPeers {
break
}
diff --git a/core/test/transport_test.go b/core/test/transport_test.go
index 9bfc12b..1d4b53d 100644
--- a/core/test/transport_test.go
+++ b/core/test/transport_test.go
@@ -32,21 +32,21 @@ import (
)
type testPeer struct {
- vID types.ValidatorID
+ nID types.NodeID
trans TransportClient
recv <-chan *TransportEnvelope
expectedEchoHash common.Hash
echoBlock *types.Block
myBlock *types.Block
myBlockSentTime time.Time
- blocks map[types.ValidatorID]*types.Block
+ blocks map[types.NodeID]*types.Block
blocksReceiveTime map[common.Hash]time.Time
}
type testPeerServer struct {
trans TransportServer
recv chan *TransportEnvelope
- peerBlocks map[types.ValidatorID]*types.Block
+ peerBlocks map[types.NodeID]*types.Block
}
type testMarshaller struct{}
@@ -88,7 +88,7 @@ type TransportTestSuite struct {
func (s *TransportTestSuite) baseTest(
server *testPeerServer,
- peers map[types.ValidatorID]*testPeer,
+ peers map[types.NodeID]*testPeer,
delay time.Duration) {
var (
@@ -98,11 +98,11 @@ func (s *TransportTestSuite) baseTest(
// For each peers, do following stuffs:
// - broadcast 1 block.
- // - report one random block to server, along with its validator ID.
+ // - report one random block to server, along with its node ID.
// Server would echo the random block back to the peer.
handleServer := func(server *testPeerServer) {
defer wg.Done()
- server.peerBlocks = make(map[types.ValidatorID]*types.Block)
+ server.peerBlocks = make(map[types.NodeID]*types.Block)
for {
select {
case e := <-server.recv:
@@ -123,14 +123,14 @@ func (s *TransportTestSuite) baseTest(
}
handlePeer := func(peer *testPeer) {
defer wg.Done()
- peer.blocks = make(map[types.ValidatorID]*types.Block)
+ peer.blocks = make(map[types.NodeID]*types.Block)
peer.blocksReceiveTime = make(map[common.Hash]time.Time)
for {
select {
case e := <-peer.recv:
switch v := e.Msg.(type) {
case *types.Block:
- if v.ProposerID == peer.vID {
+ if v.ProposerID == peer.nID {
req.Equal(e.PeerType, TransportPeerServer)
peer.echoBlock = v
} else {
@@ -150,11 +150,11 @@ func (s *TransportTestSuite) baseTest(
}
wg.Add(len(peers) + 1)
go handleServer(server)
- for vID, peer := range peers {
+ for nID, peer := range peers {
go handlePeer(peer)
// Broadcast a block.
peer.myBlock = &types.Block{
- ProposerID: vID,
+ ProposerID: nID,
Hash: common.NewRandomHash(),
}
peer.myBlockSentTime = time.Now()
@@ -162,28 +162,28 @@ func (s *TransportTestSuite) baseTest(
// Report a block to server.
peer.expectedEchoHash = common.NewRandomHash()
peer.trans.Report(&types.Block{
- ProposerID: vID,
+ ProposerID: nID,
Hash: peer.expectedEchoHash,
})
}
wg.Wait()
// Make sure each sent block is received.
- for vID, peer := range peers {
+ for nID, peer := range peers {
req.NotNil(peer.echoBlock)
req.Equal(peer.echoBlock.Hash, peer.expectedEchoHash)
- for otherVID, otherPeer := range peers {
- if vID == otherVID {
+ for othernID, otherPeer := range peers {
+ if nID == othernID {
continue
}
req.Equal(
peer.myBlock.Hash,
- otherPeer.blocks[peer.vID].Hash)
+ otherPeer.blocks[peer.nID].Hash)
}
}
// Make sure the latency is expected.
- for vID, peer := range peers {
- for otherVID, otherPeer := range peers {
- if otherVID == vID {
+ for nID, peer := range peers {
+ for othernID, otherPeer := range peers {
+ if othernID == nID {
continue
}
req.True(otherPeer.blocksReceiveTime[peer.myBlock.Hash].Sub(
@@ -196,8 +196,8 @@ func (s *TransportTestSuite) TestFake() {
var (
peerCount = 13
req = s.Require()
- peers = make(map[types.ValidatorID]*testPeer)
- vIDs = GenerateRandomValidatorIDs(peerCount)
+ peers = make(map[types.NodeID]*testPeer)
+ nIDs = GenerateRandomNodeIDs(peerCount)
err error
wg sync.WaitGroup
latency = &FixedLatencyModel{Latency: 300}
@@ -207,13 +207,13 @@ func (s *TransportTestSuite) TestFake() {
server.recv, err = server.trans.Host()
req.Nil(err)
// Setup Peers
- wg.Add(len(vIDs))
- for _, vID := range vIDs {
+ wg.Add(len(nIDs))
+ for _, nID := range nIDs {
peer := &testPeer{
- vID: vID,
- trans: NewFakeTransportClient(vID, latency),
+ nID: nID,
+ trans: NewFakeTransportClient(nID, latency),
}
- peers[vID] = peer
+ peers[nID] = peer
go func() {
defer wg.Done()
recv, err := peer.trans.Join(server.recv)
@@ -236,8 +236,8 @@ func (s *TransportTestSuite) TestTCPLocal() {
var (
peerCount = 25
req = s.Require()
- peers = make(map[types.ValidatorID]*testPeer)
- vIDs = GenerateRandomValidatorIDs(peerCount)
+ peers = make(map[types.NodeID]*testPeer)
+ nIDs = GenerateRandomNodeIDs(peerCount)
err error
wg sync.WaitGroup
latency = &FixedLatencyModel{Latency: 300}
@@ -250,13 +250,13 @@ func (s *TransportTestSuite) TestTCPLocal() {
server.recv, err = server.trans.Host()
req.Nil(err)
// Setup Peers
- wg.Add(len(vIDs))
- for _, vID := range vIDs {
+ wg.Add(len(nIDs))
+ for _, nID := range nIDs {
peer := &testPeer{
- vID: vID,
- trans: NewTCPTransportClient(vID, latency, &testMarshaller{}, true),
+ nID: nID,
+ trans: NewTCPTransportClient(nID, latency, &testMarshaller{}, true),
}
- peers[vID] = peer
+ peers[nID] = peer
go func() {
defer wg.Done()
diff --git a/core/test/utils.go b/core/test/utils.go
index 138e8a1..887ef14 100644
--- a/core/test/utils.go
+++ b/core/test/utils.go
@@ -34,11 +34,11 @@ func stableRandomHash(block *types.Block) (common.Hash, error) {
return common.NewRandomHash(), nil
}
-// GenerateRandomValidatorIDs generates randomly a slices of types.ValidatorID.
-func GenerateRandomValidatorIDs(validatorCount int) (vIDs types.ValidatorIDs) {
- vIDs = types.ValidatorIDs{}
- for i := 0; i < validatorCount; i++ {
- vIDs = append(vIDs, types.ValidatorID{Hash: common.NewRandomHash()})
+// GenerateRandomNodeIDs generates randomly a slices of types.NodeID.
+func GenerateRandomNodeIDs(nodeCount int) (nIDs types.NodeIDs) {
+ nIDs = types.NodeIDs{}
+ for i := 0; i < nodeCount; i++ {
+ nIDs = append(nIDs, types.NodeID{Hash: common.NewRandomHash()})
}
return
}
diff --git a/core/total-ordering_test.go b/core/total-ordering_test.go
index d6805b8..c9dd5a5 100644
--- a/core/total-ordering_test.go
+++ b/core/total-ordering_test.go
@@ -34,7 +34,7 @@ type TotalOrderingTestSuite struct {
}
func (s *TotalOrderingTestSuite) genGenesisBlock(
- vIDs types.ValidatorIDs,
+ vIDs types.NodeIDs,
chainID uint32,
acks common.Hashes) *types.Block {
@@ -77,9 +77,9 @@ func (s *TotalOrderingTestSuite) TestBlockRelation() {
//
// The DAG used below is:
// A <- B <- C
- validators := test.GenerateRandomValidatorIDs(5)
- vID := validators[0]
- blockA := s.genGenesisBlock(validators, 0, common.Hashes{})
+ nodes := test.GenerateRandomNodeIDs(5)
+ vID := nodes[0]
+ blockA := s.genGenesisBlock(nodes, 0, common.Hashes{})
blockB := &types.Block{
ProposerID: vID,
ParentHash: blockA.Hash,
@@ -101,7 +101,7 @@ func (s *TotalOrderingTestSuite) TestBlockRelation() {
Acks: common.NewSortedHashes(common.Hashes{blockB.Hash}),
}
- to := newTotalOrdering(1, 3, uint32(len(validators)))
+ to := newTotalOrdering(1, 3, uint32(len(nodes)))
s.checkNotDeliver(to, blockA)
s.checkNotDeliver(to, blockB)
s.checkNotDeliver(to, blockC)
@@ -204,15 +204,15 @@ func (s *TotalOrderingTestSuite) TestGrade() {
// This test case just fake some internal structure used
// when performing total ordering.
var (
- validators = test.GenerateRandomValidatorIDs(5)
- cache = newTotalOrderingObjectCache(5)
- dirtyValidators = []int{0, 1, 2, 3, 4}
+ nodes = test.GenerateRandomNodeIDs(5)
+ cache = newTotalOrderingObjectCache(5)
+ dirtyNodes = []int{0, 1, 2, 3, 4}
)
- ansLength := uint64(len(map[types.ValidatorID]struct{}{
- validators[0]: struct{}{},
- validators[1]: struct{}{},
- validators[2]: struct{}{},
- validators[3]: struct{}{},
+ ansLength := uint64(len(map[types.NodeID]struct{}{
+ nodes[0]: struct{}{},
+ nodes[1]: struct{}{},
+ nodes[2]: struct{}{},
+ nodes[3]: struct{}{},
}))
candidate1 := newTotalOrderingCandidateInfo(common.Hash{}, cache)
candidate1.cachedHeightVector = []uint64{
@@ -225,29 +225,29 @@ func (s *TotalOrderingTestSuite) TestGrade() {
1, 1, infinity, infinity, infinity}
candidate2.updateWinRecord(
- 0, candidate1, dirtyValidators, cache)
+ 0, candidate1, dirtyNodes, cache)
s.Equal(candidate2.winRecords[0].grade(5, 3, ansLength), 1)
candidate1.updateWinRecord(
- 1, candidate2, dirtyValidators, cache)
+ 1, candidate2, dirtyNodes, cache)
s.Equal(candidate1.winRecords[1].grade(5, 3, ansLength), 0)
candidate2.updateWinRecord(
- 2, candidate3, dirtyValidators, cache)
+ 2, candidate3, dirtyNodes, cache)
s.Equal(candidate2.winRecords[2].grade(5, 3, ansLength), -1)
candidate3.updateWinRecord(
- 1, candidate2, dirtyValidators, cache)
+ 1, candidate2, dirtyNodes, cache)
s.Equal(candidate3.winRecords[1].grade(5, 3, ansLength), 0)
}
func (s *TotalOrderingTestSuite) TestCycleDetection() {
// Make sure we don't get hang by cycle from
// block's acks.
- validators := test.GenerateRandomValidatorIDs(5)
+ nodes := test.GenerateRandomNodeIDs(5)
// create blocks with cycles in acking relation.
cycledHash := common.NewRandomHash()
- b00 := s.genGenesisBlock(validators, 0, common.Hashes{cycledHash})
+ b00 := s.genGenesisBlock(nodes, 0, common.Hashes{cycledHash})
b01 := &types.Block{
- ProposerID: validators[0],
+ ProposerID: nodes[0],
ParentHash: b00.Hash,
Hash: common.NewRandomHash(),
Position: types.Position{
@@ -257,7 +257,7 @@ func (s *TotalOrderingTestSuite) TestCycleDetection() {
Acks: common.NewSortedHashes(common.Hashes{b00.Hash}),
}
b02 := &types.Block{
- ProposerID: validators[0],
+ ProposerID: nodes[0],
ParentHash: b01.Hash,
Hash: common.NewRandomHash(),
Position: types.Position{
@@ -267,7 +267,7 @@ func (s *TotalOrderingTestSuite) TestCycleDetection() {
Acks: common.NewSortedHashes(common.Hashes{b01.Hash}),
}
b03 := &types.Block{
- ProposerID: validators[0],
+ ProposerID: nodes[0],
ParentHash: b02.Hash,
Hash: cycledHash,
Position: types.Position{
@@ -278,11 +278,11 @@ func (s *TotalOrderingTestSuite) TestCycleDetection() {
}
// Create a block acks self.
- b10 := s.genGenesisBlock(validators, 1, common.Hashes{})
+ b10 := s.genGenesisBlock(nodes, 1, common.Hashes{})
b10.Acks = append(b10.Acks, b10.Hash)
// Make sure we won't hang when cycle exists.
- to := newTotalOrdering(1, 3, uint32(len(validators)))
+ to := newTotalOrdering(1, 3, uint32(len(nodes)))
s.checkNotDeliver(to, b00)
s.checkNotDeliver(to, b01)
s.checkNotDeliver(to, b02)
@@ -294,12 +294,12 @@ func (s *TotalOrderingTestSuite) TestCycleDetection() {
}
func (s *TotalOrderingTestSuite) TestNotValidDAGDetection() {
- validators := test.GenerateRandomValidatorIDs(5)
- to := newTotalOrdering(1, 3, uint32(len(validators)))
+ nodes := test.GenerateRandomNodeIDs(5)
+ to := newTotalOrdering(1, 3, uint32(len(nodes)))
- b00 := s.genGenesisBlock(validators, 0, common.Hashes{})
+ b00 := s.genGenesisBlock(nodes, 0, common.Hashes{})
b01 := &types.Block{
- ProposerID: validators[0],
+ ProposerID: nodes[0],
ParentHash: b00.Hash,
Position: types.Position{
Height: 1,
@@ -326,8 +326,8 @@ func (s *TotalOrderingTestSuite) TestEarlyDeliver() {
// A B
// Even when B is not received, A should
// be able to be delivered.
- validators := test.GenerateRandomValidatorIDs(5)
- to := newTotalOrdering(2, 3, uint32(len(validators)))
+ nodes := test.GenerateRandomNodeIDs(5)
+ to := newTotalOrdering(2, 3, uint32(len(nodes)))
genNextBlock := func(b *types.Block) *types.Block {
return &types.Block{
ProposerID: b.ProposerID,
@@ -341,19 +341,19 @@ func (s *TotalOrderingTestSuite) TestEarlyDeliver() {
}
}
- b00 := s.genGenesisBlock(validators, 0, common.Hashes{})
+ b00 := s.genGenesisBlock(nodes, 0, common.Hashes{})
b01 := genNextBlock(b00)
b02 := genNextBlock(b01)
- b10 := s.genGenesisBlock(validators, 1, common.Hashes{b00.Hash})
+ b10 := s.genGenesisBlock(nodes, 1, common.Hashes{b00.Hash})
b11 := genNextBlock(b10)
b12 := genNextBlock(b11)
- b20 := s.genGenesisBlock(validators, 2, common.Hashes{b00.Hash})
+ b20 := s.genGenesisBlock(nodes, 2, common.Hashes{b00.Hash})
b21 := genNextBlock(b20)
b22 := genNextBlock(b21)
- b30 := s.genGenesisBlock(validators, 3, common.Hashes{b00.Hash})
+ b30 := s.genGenesisBlock(nodes, 3, common.Hashes{b00.Hash})
b31 := genNextBlock(b30)
b32 := genNextBlock(b31)
@@ -431,16 +431,16 @@ func (s *TotalOrderingTestSuite) TestEarlyDeliver() {
func (s *TotalOrderingTestSuite) TestBasicCaseForK2() {
// It's a handcrafted test case.
- validators := test.GenerateRandomValidatorIDs(5)
- to := newTotalOrdering(2, 3, uint32(len(validators)))
+ nodes := test.GenerateRandomNodeIDs(5)
+ to := newTotalOrdering(2, 3, uint32(len(nodes)))
// Setup blocks.
- b00 := s.genGenesisBlock(validators, 0, common.Hashes{})
- b10 := s.genGenesisBlock(validators, 1, common.Hashes{})
- b20 := s.genGenesisBlock(validators, 2, common.Hashes{b10.Hash})
- b30 := s.genGenesisBlock(validators, 3, common.Hashes{b20.Hash})
- b40 := s.genGenesisBlock(validators, 4, common.Hashes{})
+ b00 := s.genGenesisBlock(nodes, 0, common.Hashes{})
+ b10 := s.genGenesisBlock(nodes, 1, common.Hashes{})
+ b20 := s.genGenesisBlock(nodes, 2, common.Hashes{b10.Hash})
+ b30 := s.genGenesisBlock(nodes, 3, common.Hashes{b20.Hash})
+ b40 := s.genGenesisBlock(nodes, 4, common.Hashes{})
b11 := &types.Block{
- ProposerID: validators[1],
+ ProposerID: nodes[1],
ParentHash: b10.Hash,
Hash: common.NewRandomHash(),
Position: types.Position{
@@ -450,7 +450,7 @@ func (s *TotalOrderingTestSuite) TestBasicCaseForK2() {
Acks: common.NewSortedHashes(common.Hashes{b10.Hash, b00.Hash}),
}
b01 := &types.Block{
- ProposerID: validators[0],
+ ProposerID: nodes[0],
ParentHash: b00.Hash,
Hash: common.NewRandomHash(),
Position: types.Position{
@@ -460,7 +460,7 @@ func (s *TotalOrderingTestSuite) TestBasicCaseForK2() {
Acks: common.NewSortedHashes(common.Hashes{b00.Hash, b11.Hash}),
}
b21 := &types.Block{
- ProposerID: validators[2],
+ ProposerID: nodes[2],
ParentHash: b20.Hash,
Hash: common.NewRandomHash(),
Position: types.Position{
@@ -470,7 +470,7 @@ func (s *TotalOrderingTestSuite) TestBasicCaseForK2() {
Acks: common.NewSortedHashes(common.Hashes{b20.Hash, b01.Hash}),
}
b31 := &types.Block{
- ProposerID: validators[3],
+ ProposerID: nodes[3],
ParentHash: b30.Hash,
Hash: common.NewRandomHash(),
Position: types.Position{
@@ -480,7 +480,7 @@ func (s *TotalOrderingTestSuite) TestBasicCaseForK2() {
Acks: common.NewSortedHashes(common.Hashes{b30.Hash, b21.Hash}),
}
b02 := &types.Block{
- ProposerID: validators[0],
+ ProposerID: nodes[0],
ParentHash: b01.Hash,
Hash: common.NewRandomHash(),
Position: types.Position{
@@ -490,7 +490,7 @@ func (s *TotalOrderingTestSuite) TestBasicCaseForK2() {
Acks: common.NewSortedHashes(common.Hashes{b01.Hash, b21.Hash}),
}
b12 := &types.Block{
- ProposerID: validators[1],
+ ProposerID: nodes[1],
ParentHash: b11.Hash,
Hash: common.NewRandomHash(),
Position: types.Position{
@@ -500,7 +500,7 @@ func (s *TotalOrderingTestSuite) TestBasicCaseForK2() {
Acks: common.NewSortedHashes(common.Hashes{b11.Hash, b21.Hash}),
}
b32 := &types.Block{
- ProposerID: validators[3],
+ ProposerID: nodes[3],
ParentHash: b31.Hash,
Hash: common.NewRandomHash(),
Position: types.Position{
@@ -510,7 +510,7 @@ func (s *TotalOrderingTestSuite) TestBasicCaseForK2() {
Acks: common.NewSortedHashes(common.Hashes{b31.Hash}),
}
b22 := &types.Block{
- ProposerID: validators[2],
+ ProposerID: nodes[2],
ParentHash: b21.Hash,
Hash: common.NewRandomHash(),
Position: types.Position{
@@ -520,7 +520,7 @@ func (s *TotalOrderingTestSuite) TestBasicCaseForK2() {
Acks: common.NewSortedHashes(common.Hashes{b21.Hash, b32.Hash}),
}
b23 := &types.Block{
- ProposerID: validators[2],
+ ProposerID: nodes[2],
ParentHash: b22.Hash,
Hash: common.NewRandomHash(),
Position: types.Position{
@@ -530,7 +530,7 @@ func (s *TotalOrderingTestSuite) TestBasicCaseForK2() {
Acks: common.NewSortedHashes(common.Hashes{b22.Hash}),
}
b03 := &types.Block{
- ProposerID: validators[0],
+ ProposerID: nodes[0],
ParentHash: b02.Hash,
Hash: common.NewRandomHash(),
Position: types.Position{
@@ -540,7 +540,7 @@ func (s *TotalOrderingTestSuite) TestBasicCaseForK2() {
Acks: common.NewSortedHashes(common.Hashes{b02.Hash, b22.Hash}),
}
b13 := &types.Block{
- ProposerID: validators[1],
+ ProposerID: nodes[1],
ParentHash: b12.Hash,
Hash: common.NewRandomHash(),
Position: types.Position{
@@ -550,7 +550,7 @@ func (s *TotalOrderingTestSuite) TestBasicCaseForK2() {
Acks: common.NewSortedHashes(common.Hashes{b12.Hash, b22.Hash}),
}
b14 := &types.Block{
- ProposerID: validators[1],
+ ProposerID: nodes[1],
ParentHash: b13.Hash,
Hash: common.NewRandomHash(),
Position: types.Position{
@@ -560,7 +560,7 @@ func (s *TotalOrderingTestSuite) TestBasicCaseForK2() {
Acks: common.NewSortedHashes(common.Hashes{b13.Hash}),
}
b41 := &types.Block{
- ProposerID: validators[4],
+ ProposerID: nodes[4],
ParentHash: b40.Hash,
Hash: common.NewRandomHash(),
Position: types.Position{
@@ -570,7 +570,7 @@ func (s *TotalOrderingTestSuite) TestBasicCaseForK2() {
Acks: common.NewSortedHashes(common.Hashes{b40.Hash}),
}
b42 := &types.Block{
- ProposerID: validators[4],
+ ProposerID: nodes[4],
ParentHash: b41.Hash,
Hash: common.NewRandomHash(),
Position: types.Position{
@@ -766,17 +766,17 @@ func (s *TotalOrderingTestSuite) TestBasicCaseForK0() {
// v v v v
// o o o <- o Height: 0
var (
- req = s.Require()
- validators = test.GenerateRandomValidatorIDs(5)
- to = newTotalOrdering(0, 3, uint32(len(validators)))
+ req = s.Require()
+ nodes = test.GenerateRandomNodeIDs(5)
+ to = newTotalOrdering(0, 3, uint32(len(nodes)))
)
// Setup blocks.
- b00 := s.genGenesisBlock(validators, 0, common.Hashes{})
- b10 := s.genGenesisBlock(validators, 1, common.Hashes{})
- b20 := s.genGenesisBlock(validators, 2, common.Hashes{})
- b30 := s.genGenesisBlock(validators, 3, common.Hashes{b20.Hash})
+ b00 := s.genGenesisBlock(nodes, 0, common.Hashes{})
+ b10 := s.genGenesisBlock(nodes, 1, common.Hashes{})
+ b20 := s.genGenesisBlock(nodes, 2, common.Hashes{})
+ b30 := s.genGenesisBlock(nodes, 3, common.Hashes{b20.Hash})
b01 := &types.Block{
- ProposerID: validators[0],
+ ProposerID: nodes[0],
ParentHash: b00.Hash,
Hash: common.NewRandomHash(),
Position: types.Position{
@@ -786,7 +786,7 @@ func (s *TotalOrderingTestSuite) TestBasicCaseForK0() {
Acks: common.NewSortedHashes(common.Hashes{b00.Hash, b10.Hash}),
}
b11 := &types.Block{
- ProposerID: validators[1],
+ ProposerID: nodes[1],
ParentHash: b10.Hash,
Hash: common.NewRandomHash(),
Position: types.Position{
@@ -796,7 +796,7 @@ func (s *TotalOrderingTestSuite) TestBasicCaseForK0() {
Acks: common.NewSortedHashes(common.Hashes{b10.Hash, b20.Hash}),
}
b21 := &types.Block{
- ProposerID: validators[2],
+ ProposerID: nodes[2],
ParentHash: b20.Hash,
Hash: common.NewRandomHash(),
Position: types.Position{
@@ -806,7 +806,7 @@ func (s *TotalOrderingTestSuite) TestBasicCaseForK0() {
Acks: common.NewSortedHashes(common.Hashes{b20.Hash}),
}
b31 := &types.Block{
- ProposerID: validators[3],
+ ProposerID: nodes[3],
ParentHash: b30.Hash,
Hash: common.NewRandomHash(),
Position: types.Position{
@@ -815,7 +815,7 @@ func (s *TotalOrderingTestSuite) TestBasicCaseForK0() {
},
Acks: common.NewSortedHashes(common.Hashes{b21.Hash, b30.Hash}),
}
- b40 := s.genGenesisBlock(validators, 4, common.Hashes{b31.Hash})
+ b40 := s.genGenesisBlock(nodes, 4, common.Hashes{b31.Hash})
s.checkNotDeliver(to, b00)
s.checkNotDeliver(to, b10)
@@ -864,8 +864,8 @@ func (s *TotalOrderingTestSuite) TestBasicCaseForK0() {
}
func (s *TotalOrderingTestSuite) baseTestRandomlyGeneratedBlocks(
- totalOrderingConstructor func(types.ValidatorIDs) *totalOrdering,
- validatorCount, blockCount int,
+ totalOrderingConstructor func(types.NodeIDs) *totalOrdering,
+ nodeCount, blockCount int,
ackingCountGenerator func() int,
repeat int) {
@@ -878,10 +878,10 @@ func (s *TotalOrderingTestSuite) baseTestRandomlyGeneratedBlocks(
db, err := blockdb.NewMemBackedBlockDB()
req.Nil(err)
- validators, err := gen.Generate(
- validatorCount, blockCount, ackingCountGenerator, db)
+ nodes, err := gen.Generate(
+ nodeCount, blockCount, ackingCountGenerator, db)
req.Nil(err)
- req.Len(validators, validatorCount)
+ req.Len(nodes, nodeCount)
iter, err := db.GetAll()
req.Nil(err)
// Setup a revealer that would reveal blocks forming
@@ -894,7 +894,7 @@ func (s *TotalOrderingTestSuite) baseTestRandomlyGeneratedBlocks(
revealed := ""
ordered := ""
revealer.Reset()
- to := totalOrderingConstructor(validators)
+ to := totalOrderingConstructor(nodes)
for {
// Reveal next block.
b, err := revealer.Next()
@@ -937,44 +937,44 @@ func (s *TotalOrderingTestSuite) baseTestRandomlyGeneratedBlocks(
func (s *TotalOrderingTestSuite) TestRandomlyGeneratedBlocks() {
var (
- validatorCount = 13
- blockCount = 50
- phi uint64 = 10
- repeat = 8
+ nodeCount = 13
+ blockCount = 50
+ phi uint64 = 10
+ repeat = 8
)
ackingCountGenerators := []func() int{
nil, // Acking frequency with normal distribution.
- test.MaxAckingCountGenerator(0), // Low acking frequency.
- test.MaxAckingCountGenerator(validatorCount), // High acking frequency.
+ test.MaxAckingCountGenerator(0), // Low acking frequency.
+ test.MaxAckingCountGenerator(nodeCount), // High acking frequency.
}
// Test based on different acking frequency.
for _, gen := range ackingCountGenerators {
// Test for K=0.
- constructor := func(validators types.ValidatorIDs) *totalOrdering {
- return newTotalOrdering(0, phi, uint32(len(validators)))
+ constructor := func(nodes types.NodeIDs) *totalOrdering {
+ return newTotalOrdering(0, phi, uint32(len(nodes)))
}
s.baseTestRandomlyGeneratedBlocks(
- constructor, validatorCount, blockCount, gen, repeat)
+ constructor, nodeCount, blockCount, gen, repeat)
// Test for K=1,
- constructor = func(validators types.ValidatorIDs) *totalOrdering {
- return newTotalOrdering(1, phi, uint32(len(validators)))
+ constructor = func(nodes types.NodeIDs) *totalOrdering {
+ return newTotalOrdering(1, phi, uint32(len(nodes)))
}
s.baseTestRandomlyGeneratedBlocks(
- constructor, validatorCount, blockCount, gen, repeat)
+ constructor, nodeCount, blockCount, gen, repeat)
// Test for K=2,
- constructor = func(validators types.ValidatorIDs) *totalOrdering {
- return newTotalOrdering(2, phi, uint32(len(validators)))
+ constructor = func(nodes types.NodeIDs) *totalOrdering {
+ return newTotalOrdering(2, phi, uint32(len(nodes)))
}
s.baseTestRandomlyGeneratedBlocks(
- constructor, validatorCount, blockCount, gen, repeat)
+ constructor, nodeCount, blockCount, gen, repeat)
// Test for K=3,
- constructor = func(validators types.ValidatorIDs) *totalOrdering {
- return newTotalOrdering(3, phi, uint32(len(validators)))
+ constructor = func(nodes types.NodeIDs) *totalOrdering {
+ return newTotalOrdering(3, phi, uint32(len(nodes)))
}
s.baseTestRandomlyGeneratedBlocks(
- constructor, validatorCount, blockCount, gen, repeat)
+ constructor, nodeCount, blockCount, gen, repeat)
}
}
diff --git a/core/types/block.go b/core/types/block.go
index 61cb458..949876f 100644
--- a/core/types/block.go
+++ b/core/types/block.go
@@ -54,7 +54,7 @@ func NewBlock() (b *Block) {
// Block represents a single event broadcasted on the network.
type Block struct {
- ProposerID ValidatorID `json:"proposer_id"`
+ ProposerID NodeID `json:"proposer_id"`
ParentHash common.Hash `json:"parent_hash"`
Hash common.Hash `json:"hash"`
Position Position `json:"position"`
diff --git a/core/types/dkg.go b/core/types/dkg.go
index 7fb686c..78f4da6 100644
--- a/core/types/dkg.go
+++ b/core/types/dkg.go
@@ -24,7 +24,7 @@ import (
// DKGPrivateShare describe a secret share in DKG protocol.
type DKGPrivateShare struct {
- ProposerID ValidatorID `json:"proposer_id"`
+ ProposerID NodeID `json:"proposer_id"`
Round uint64 `json:"round"`
PrivateShare dkg.PrivateKey `json:"private_share"`
Signature crypto.Signature `json:"signature"`
@@ -32,7 +32,7 @@ type DKGPrivateShare struct {
// DKGMasterPublicKey decrtibe a master public key in DKG protocol.
type DKGMasterPublicKey struct {
- ProposerID ValidatorID `json:"proposer_id"`
+ ProposerID NodeID `json:"proposer_id"`
Round uint64 `json:"round"`
DKGID dkg.ID `json:"dkg_id"`
PublicKeyShares dkg.PublicKeyShares `json:"public_key_shares"`
@@ -41,7 +41,7 @@ type DKGMasterPublicKey struct {
// DKGComplaint describe a complaint in DKG protocol.
type DKGComplaint struct {
- ProposerID ValidatorID `json:"proposer_id"`
+ ProposerID NodeID `json:"proposer_id"`
Round uint64 `json:"round"`
PrivateShare DKGPrivateShare `json:"private_share"`
Signature crypto.Signature `json:"signature"`
@@ -49,7 +49,7 @@ type DKGComplaint struct {
// DKGPartialSignature describe a partial signature in DKG protocol.
type DKGPartialSignature struct {
- ProposerID ValidatorID `json:"proposerID"`
+ ProposerID NodeID `json:"proposerID"`
Round uint64 `json:"round"`
PartialSignature dkg.PartialSignature `json:"partial_signature"`
Signature crypto.Signature `json:"signature"`
diff --git a/core/types/validator.go b/core/types/node.go
index 5151a6d..8a856de 100644
--- a/core/types/validator.go
+++ b/core/types/node.go
@@ -24,33 +24,33 @@ import (
"github.com/dexon-foundation/dexon-consensus-core/crypto"
)
-// ValidatorID is the ID type for validators.
-type ValidatorID struct {
+// NodeID is the ID type for nodes.
+type NodeID struct {
common.Hash
}
-// NewValidatorID returns a ValidatorID with Hash set to the hash value of
+// NewNodeID returns a NodeID with Hash set to the hash value of
// public key.
-func NewValidatorID(pubKey crypto.PublicKey) ValidatorID {
- return ValidatorID{Hash: crypto.Keccak256Hash(pubKey.Bytes())}
+func NewNodeID(pubKey crypto.PublicKey) NodeID {
+ return NodeID{Hash: crypto.Keccak256Hash(pubKey.Bytes())}
}
-// Equal checks if the hash representation is the same ValidatorID.
-func (v ValidatorID) Equal(hash common.Hash) bool {
+// Equal checks if the hash representation is the same NodeID.
+func (v NodeID) Equal(hash common.Hash) bool {
return v.Hash == hash
}
-// ValidatorIDs implements sort.Interface for ValidatorID.
-type ValidatorIDs []ValidatorID
+// NodeIDs implements sort.Interface for NodeID.
+type NodeIDs []NodeID
-func (v ValidatorIDs) Len() int {
+func (v NodeIDs) Len() int {
return len(v)
}
-func (v ValidatorIDs) Less(i int, j int) bool {
+func (v NodeIDs) Less(i int, j int) bool {
return bytes.Compare([]byte(v[i].Hash[:]), []byte(v[j].Hash[:])) == -1
}
-func (v ValidatorIDs) Swap(i int, j int) {
+func (v NodeIDs) Swap(i int, j int) {
v[i], v[j] = v[j], v[i]
}
diff --git a/core/types/vote.go b/core/types/vote.go
index bae8f7d..e92aa67 100644
--- a/core/types/vote.go
+++ b/core/types/vote.go
@@ -38,7 +38,7 @@ const (
// Vote is the vote structure defined in Crypto Shuffle Algorithm.
type Vote struct {
- ProposerID ValidatorID `json:"proposer_id"`
+ ProposerID NodeID `json:"proposer_id"`
Type VoteType `json:"type"`
BlockHash common.Hash `json:"block_hash"`
Period uint64 `json:"period"`
diff --git a/core/types/witness.go b/core/types/witness.go
index 1e432f6..349c1ab 100644
--- a/core/types/witness.go
+++ b/core/types/witness.go
@@ -28,7 +28,7 @@ import (
// WitnessAck represents the acking to the compaction chain.
type WitnessAck struct {
- ProposerID ValidatorID `json:"proposer_id"`
+ ProposerID NodeID `json:"proposer_id"`
WitnessBlockHash common.Hash `json:"witness_block_hash"`
Hash common.Hash `json:"hash"`
// WitnessSignature is the signature of the hash value of BlockWitness.