aboutsummaryrefslogblamecommitdiffstats
path: root/cmd/swarm/run_test.go
blob: 680d238d0601446e0e4bc289645191f2016e7656 (plain) (tree)
1
                                         

















                                                                       
                 
                      
              



                   
              

                       

                 



                                             
                                                  





                                                           

                                                                  

 

                                                           










                                                                        










                                                        
                                                    

                                           

































                                                                               
 














                                                          
                                      




































                                                                                                             



































                                                                                     
                      








                                    



                                              
                                                                                             
                     
                            

                                    
                              




                                
                                                                                                                            



                            




                                                                                   


                            









                                                                              
                                             


                            

                            



                                  
                                           






                                           
                                                     







                                          










                                                                                






















                                                                                                        

                                   


                   


                                                      




                                                                                      
 
                       
                                             


                            

                            



                                  
                                           


                                          
                                


                                                         
                                                     







                                          










                                                                                






















                                                                                                        
                                   
                                   








                               






















                                                                            
         






























































                                                                                                                         
         
 
// Copyright 2017 The go-ethereum Authors
// This file is part of go-ethereum.
//
// go-ethereum is free software: you can redistribute it and/or modify
// it under the terms of the GNU General Public License as published by
// the Free Software Foundation, either version 3 of the License, or
// (at your option) any later version.
//
// go-ethereum is distributed in the hope that it will be useful,
// but WITHOUT ANY WARRANTY; without even the implied warranty of
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
// GNU General Public License for more details.
//
// You should have received a copy of the GNU General Public License
// along with go-ethereum. If not, see <http://www.gnu.org/licenses/>.

package main

import (
    "context"
    "crypto/ecdsa"
    "flag"
    "fmt"
    "io/ioutil"
    "net"
    "os"
    "path"
    "path/filepath"
    "runtime"
    "sync"
    "syscall"
    "testing"
    "time"

    "github.com/docker/docker/pkg/reexec"
    "github.com/ethereum/go-ethereum/accounts"
    "github.com/ethereum/go-ethereum/accounts/keystore"
    "github.com/ethereum/go-ethereum/internal/cmdtest"
    "github.com/ethereum/go-ethereum/node"
    "github.com/ethereum/go-ethereum/p2p"
    "github.com/ethereum/go-ethereum/rpc"
    "github.com/ethereum/go-ethereum/swarm"
    "github.com/ethereum/go-ethereum/swarm/api"
    swarmhttp "github.com/ethereum/go-ethereum/swarm/api/http"
)

var loglevel = flag.Int("loglevel", 3, "verbosity of logs")

func init() {
    // Run the app if we've been exec'd as "swarm-test" in runSwarm.
    reexec.Register("swarm-test", func() {
        if err := app.Run(os.Args); err != nil {
            fmt.Fprintln(os.Stderr, err)
            os.Exit(1)
        }
        os.Exit(0)
    })
}

const clusterSize = 3

var clusteronce sync.Once
var cluster *testCluster

func initCluster(t *testing.T) {
    clusteronce.Do(func() {
        cluster = newTestCluster(t, clusterSize)
    })
}

func serverFunc(api *api.API) swarmhttp.TestServer {
    return swarmhttp.NewServer(api, "")
}
func TestMain(m *testing.M) {
    // check if we have been reexec'd
    if reexec.Init() {
        return
    }
    os.Exit(m.Run())
}

func runSwarm(t *testing.T, args ...string) *cmdtest.TestCmd {
    tt := cmdtest.NewTestCmd(t, nil)

    // Boot "swarm". This actually runs the test binary but the TestMain
    // function will prevent any tests from running.
    tt.Run("swarm-test", args...)

    return tt
}

type testCluster struct {
    Nodes  []*testNode
    TmpDir string
}

// newTestCluster starts a test swarm cluster of the given size.
//
// A temporary directory is created and each node gets a data directory inside
// it.
//
// Each node listens on 127.0.0.1 with random ports for both the HTTP and p2p
// ports (assigned by first listening on 127.0.0.1:0 and then passing the ports
// as flags).
//
// When starting more than one node, they are connected together using the
// admin SetPeer RPC method.

func newTestCluster(t *testing.T, size int) *testCluster {
    cluster := &testCluster{}
    defer func() {
        if t.Failed() {
            cluster.Shutdown()
        }
    }()

    tmpdir, err := ioutil.TempDir("", "swarm-test")
    if err != nil {
        t.Fatal(err)
    }
    cluster.TmpDir = tmpdir

    // start the nodes
    cluster.StartNewNodes(t, size)

    if size == 1 {
        return cluster
    }

    // connect the nodes together
    for _, node := range cluster.Nodes {
        if err := node.Client.Call(nil, "admin_addPeer", cluster.Nodes[0].Enode); err != nil {
            t.Fatal(err)
        }
    }

    // wait until all nodes have the correct number of peers
outer:
    for _, node := range cluster.Nodes {
        var peers []*p2p.PeerInfo
        for start := time.Now(); time.Since(start) < time.Minute; time.Sleep(50 * time.Millisecond) {
            if err := node.Client.Call(&peers, "admin_peers"); err != nil {
                t.Fatal(err)
            }
            if len(peers) == len(cluster.Nodes)-1 {
                continue outer
            }
        }
        t.Fatalf("%s only has %d / %d peers", node.Name, len(peers), len(cluster.Nodes)-1)
    }

    return cluster
}

func (c *testCluster) Shutdown() {
    for _, node := range c.Nodes {
        node.Shutdown()
    }
    os.RemoveAll(c.TmpDir)
}

func (c *testCluster) Stop() {
    for _, node := range c.Nodes {
        node.Shutdown()
    }
}

func (c *testCluster) StartNewNodes(t *testing.T, size int) {
    c.Nodes = make([]*testNode, 0, size)
    for i := 0; i < size; i++ {
        dir := filepath.Join(c.TmpDir, fmt.Sprintf("swarm%02d", i))
        if err := os.Mkdir(dir, 0700); err != nil {
            t.Fatal(err)
        }

        node := newTestNode(t, dir)
        node.Name = fmt.Sprintf("swarm%02d", i)

        c.Nodes = append(c.Nodes, node)
    }
}

func (c *testCluster) StartExistingNodes(t *testing.T, size int, bzzaccount string) {
    c.Nodes = make([]*testNode, 0, size)
    for i := 0; i < size; i++ {
        dir := filepath.Join(c.TmpDir, fmt.Sprintf("swarm%02d", i))
        node := existingTestNode(t, dir, bzzaccount)
        node.Name = fmt.Sprintf("swarm%02d", i)

        c.Nodes = append(c.Nodes, node)
    }
}

func (c *testCluster) Cleanup() {
    os.RemoveAll(c.TmpDir)
}

type testNode struct {
    Name       string
    Addr       string
    URL        string
    Enode      string
    Dir        string
    IpcPath    string
    PrivateKey *ecdsa.PrivateKey
    Client     *rpc.Client
    Cmd        *cmdtest.TestCmd
}

const testPassphrase = "swarm-test-passphrase"

func getTestAccount(t *testing.T, dir string) (conf *node.Config, account accounts.Account) {
    // create key
    conf = &node.Config{
        DataDir: dir,
        IPCPath: "bzzd.ipc",
        NoUSB:   true,
    }
    n, err := node.New(conf)
    if err != nil {
        t.Fatal(err)
    }
    account, err = n.AccountManager().Backends(keystore.KeyStoreType)[0].(*keystore.KeyStore).NewAccount(testPassphrase)
    if err != nil {
        t.Fatal(err)
    }

    // use a unique IPCPath when running tests on Windows
    if runtime.GOOS == "windows" {
        conf.IPCPath = fmt.Sprintf("bzzd-%s.ipc", account.Address.String())
    }

    return conf, account
}

func existingTestNode(t *testing.T, dir string, bzzaccount string) *testNode {
    conf, _ := getTestAccount(t, dir)
    node := &testNode{Dir: dir}

    // use a unique IPCPath when running tests on Windows
    if runtime.GOOS == "windows" {
        conf.IPCPath = fmt.Sprintf("bzzd-%s.ipc", bzzaccount)
    }

    // assign ports
    ports, err := getAvailableTCPPorts(2)
    if err != nil {
        t.Fatal(err)
    }
    p2pPort := ports[0]
    httpPort := ports[1]

    // start the node
    node.Cmd = runSwarm(t,
        "--port", p2pPort,
        "--nat", "extip:127.0.0.1",
        "--nodiscover",
        "--datadir", dir,
        "--ipcpath", conf.IPCPath,
        "--ens-api", "",
        "--bzzaccount", bzzaccount,
        "--bzznetworkid", "321",
        "--bzzport", httpPort,
        "--verbosity", fmt.Sprint(*loglevel),
    )
    node.Cmd.InputLine(testPassphrase)
    defer func() {
        if t.Failed() {
            node.Shutdown()
        }
    }()

    ctx, cancel := context.WithTimeout(context.Background(), 30*time.Second)
    defer cancel()

    // ensure that all ports have active listeners
    // so that the next node will not get the same
    // when calling getAvailableTCPPorts
    err = waitTCPPorts(ctx, ports...)
    if err != nil {
        t.Fatal(err)
    }

    // wait for the node to start
    for start := time.Now(); time.Since(start) < 10*time.Second; time.Sleep(50 * time.Millisecond) {
        node.Client, err = rpc.Dial(conf.IPCEndpoint())
        if err == nil {
            break
        }
    }
    if node.Client == nil {
        t.Fatal(err)
    }

    // load info
    var info swarm.Info
    if err := node.Client.Call(&info, "bzz_info"); err != nil {
        t.Fatal(err)
    }
    node.Addr = net.JoinHostPort("127.0.0.1", info.Port)
    node.URL = "http://" + node.Addr

    var nodeInfo p2p.NodeInfo
    if err := node.Client.Call(&nodeInfo, "admin_nodeInfo"); err != nil {
        t.Fatal(err)
    }
    node.Enode = nodeInfo.Enode
    node.IpcPath = conf.IPCPath
    return node
}

func newTestNode(t *testing.T, dir string) *testNode {

    conf, account := getTestAccount(t, dir)
    ks := keystore.NewKeyStore(path.Join(dir, "keystore"), 1<<18, 1)

    pk := decryptStoreAccount(ks, account.Address.Hex(), []string{testPassphrase})

    node := &testNode{Dir: dir, PrivateKey: pk}

    // assign ports
    ports, err := getAvailableTCPPorts(2)
    if err != nil {
        t.Fatal(err)
    }
    p2pPort := ports[0]
    httpPort := ports[1]

    // start the node
    node.Cmd = runSwarm(t,
        "--port", p2pPort,
        "--nat", "extip:127.0.0.1",
        "--nodiscover",
        "--datadir", dir,
        "--ipcpath", conf.IPCPath,
        "--ens-api", "",
        "--bzzaccount", account.Address.String(),
        "--bzznetworkid", "321",
        "--bzzport", httpPort,
        "--verbosity", fmt.Sprint(*loglevel),
    )
    node.Cmd.InputLine(testPassphrase)
    defer func() {
        if t.Failed() {
            node.Shutdown()
        }
    }()

    ctx, cancel := context.WithTimeout(context.Background(), 30*time.Second)
    defer cancel()

    // ensure that all ports have active listeners
    // so that the next node will not get the same
    // when calling getAvailableTCPPorts
    err = waitTCPPorts(ctx, ports...)
    if err != nil {
        t.Fatal(err)
    }

    // wait for the node to start
    for start := time.Now(); time.Since(start) < 10*time.Second; time.Sleep(50 * time.Millisecond) {
        node.Client, err = rpc.Dial(conf.IPCEndpoint())
        if err == nil {
            break
        }
    }
    if node.Client == nil {
        t.Fatal(err)
    }

    // load info
    var info swarm.Info
    if err := node.Client.Call(&info, "bzz_info"); err != nil {
        t.Fatal(err)
    }
    node.Addr = net.JoinHostPort("127.0.0.1", info.Port)
    node.URL = "http://" + node.Addr

    var nodeInfo p2p.NodeInfo
    if err := node.Client.Call(&nodeInfo, "admin_nodeInfo"); err != nil {
        t.Fatal(err)
    }
    node.Enode = nodeInfo.Enode
    node.IpcPath = conf.IPCPath
    return node
}

func (n *testNode) Shutdown() {
    if n.Cmd != nil {
        n.Cmd.Kill()
    }
}

// getAvailableTCPPorts returns a set of ports that
// nothing is listening on at the time.
//
// Function assignTCPPort cannot be called in sequence
// and guardantee that the same port will be returned in
// different calls as the listener is closed within the function,
// not after all listeners are started and selected unique
// available ports.
func getAvailableTCPPorts(count int) (ports []string, err error) {
    for i := 0; i < count; i++ {
        l, err := net.Listen("tcp", "127.0.0.1:0")
        if err != nil {
            return nil, err
        }
        // defer close in the loop to be sure the same port will not
        // be selected in the next iteration
        defer l.Close()

        _, port, err := net.SplitHostPort(l.Addr().String())
        if err != nil {
            return nil, err
        }
        ports = append(ports, port)
    }
    return ports, nil
}

// waitTCPPorts blocks until tcp connections can be
// established on all provided ports. It runs all
// ports dialers in parallel, and returns the first
// encountered error.
// See waitTCPPort also.
func waitTCPPorts(ctx context.Context, ports ...string) error {
    var err error
    // mu locks err variable that is assigned in
    // other goroutines
    var mu sync.Mutex

    // cancel is canceling all goroutines
    // when the firs error is returned
    // to prevent unnecessary waiting
    ctx, cancel := context.WithCancel(ctx)
    defer cancel()

    var wg sync.WaitGroup
    for _, port := range ports {
        wg.Add(1)
        go func(port string) {
            defer wg.Done()

            e := waitTCPPort(ctx, port)

            mu.Lock()
            defer mu.Unlock()
            if e != nil && err == nil {
                err = e
                cancel()
            }
        }(port)
    }
    wg.Wait()

    return err
}

// waitTCPPort blocks until tcp connection can be established
// ona provided port. It has a 3 minute timeout as maximum,
// to prevent long waiting, but it can be shortened with
// a provided context instance. Dialer has a 10 second timeout
// in every iteration, and connection refused error will be
// retried in 100 milliseconds periods.
func waitTCPPort(ctx context.Context, port string) error {
    ctx, cancel := context.WithTimeout(ctx, 3*time.Minute)
    defer cancel()

    for {
        c, err := (&net.Dialer{Timeout: 10 * time.Second}).DialContext(ctx, "tcp", "127.0.0.1:"+port)
        if err != nil {
            if operr, ok := err.(*net.OpError); ok {
                if syserr, ok := operr.Err.(*os.SyscallError); ok && syserr.Err == syscall.ECONNREFUSED {
                    time.Sleep(100 * time.Millisecond)
                    continue
                }
            }
            return err
        }
        return c.Close()
    }
}