aboutsummaryrefslogtreecommitdiffstats
path: root/simulation/simulation.go
blob: 978107a6d8dfe2259cbbe46d03acb118171dd29f (plain) (blame)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
// Copyright 2018 The dexon-consensus-core Authors
// This file is part of the dexon-consensus-core library.
//
// The dexon-consensus-core library is free software: you can redistribute it
// and/or modify it under the terms of the GNU Lesser General Public License as
// published by the Free Software Foundation, either version 3 of the License,
// or (at your option) any later version.
//
// The dexon-consensus-core library is distributed in the hope that it will be
// useful, but WITHOUT ANY WARRANTY; without even the implied warranty of
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU Lesser
// General Public License for more details.
//
// You should have received a copy of the GNU Lesser General Public License
// along with the dexon-consensus-core library. If not, see
// <http://www.gnu.org/licenses/>.

package simulation

import (
    "fmt"
    "sync"

    "github.com/dexon-foundation/dexon-consensus-core/crypto/eth"
    "github.com/dexon-foundation/dexon-consensus-core/simulation/config"
)

// Run starts the simulation.
func Run(configPath string, legacy bool) {
    cfg, err := config.Read(configPath)
    if err != nil {
        panic(err)
    }

    networkType := cfg.Networking.Type

    var (
        vs           []*Validator
        networkModel = &NormalNetwork{
            Sigma:         cfg.Networking.Sigma,
            Mean:          cfg.Networking.Mean,
            LossRateValue: cfg.Networking.LossRateValue,
        }
    )

    if networkType == config.NetworkTypeTCPLocal {
        lock := sync.Mutex{}
        wg := sync.WaitGroup{}
        for i := 0; i < cfg.Validator.Num; i++ {
            prv, err := eth.NewPrivateKey()
            if err != nil {
                panic(err)
            }
            wg.Add(1)
            go func() {
                network := NewTCPNetwork(true, cfg.Networking.PeerServer, networkModel)
                network.Start()
                lock.Lock()
                defer lock.Unlock()
                vs = append(vs, NewValidator(prv, eth.SigToPub, cfg.Validator, network))
                wg.Done()
            }()
        }
        wg.Wait()

        for i := 0; i < cfg.Validator.Num; i++ {
            fmt.Printf("Validator %d: %s\n", i, vs[i].ID)
            go vs[i].Run(legacy)
        }
    } else if networkType == config.NetworkTypeTCP {
        prv, err := eth.NewPrivateKey()
        if err != nil {
            panic(err)
        }
        network := NewTCPNetwork(false, cfg.Networking.PeerServer, networkModel)
        network.Start()
        v := NewValidator(prv, eth.SigToPub, cfg.Validator, network)
        go v.Run(legacy)
        vs = append(vs, v)
    }

    for _, v := range vs {
        v.Wait()
        fmt.Printf("Validator %s is shutdown\n", v.GetID())
    }

    // Do not exit when we are in TCP node, since k8s will restart the pod and
    // cause confusions.
    if networkType == config.NetworkTypeTCP {
        select {}
    }
}