| 123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328 |
- // Copyright 2018 The go-ethereum Authors
- // This file is part of the go-ethereum library.
- //
- // The go-ethereum library is free software: you can redistribute it and/or modify
- // it under the terms of the GNU Lesser General Public License as published by
- // the Free Software Foundation, either version 3 of the License, or
- // (at your option) any later version.
- //
- // The go-ethereum library is distributed in the hope that it will be useful,
- // but WITHOUT ANY WARRANTY; without even the implied warranty of
- // MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
- // GNU Lesser General Public License for more details.
- //
- // You should have received a copy of the GNU Lesser General Public License
- // along with the go-ethereum library. If not, see <http://www.gnu.org/licenses/>.
- package stream
- import (
- "context"
- "fmt"
- "os"
- "sync"
- "testing"
- "time"
- "github.com/ethereum/go-ethereum/node"
- "github.com/ethereum/go-ethereum/p2p/enode"
- "github.com/ethereum/go-ethereum/p2p/simulations/adapters"
- "github.com/ethereum/go-ethereum/swarm/log"
- "github.com/ethereum/go-ethereum/swarm/network"
- "github.com/ethereum/go-ethereum/swarm/network/simulation"
- "github.com/ethereum/go-ethereum/swarm/state"
- "github.com/ethereum/go-ethereum/swarm/storage"
- )
- //constants for random file generation
- const (
- minFileSize = 2
- maxFileSize = 40
- )
- //This test is a retrieval test for nodes.
- //A configurable number of nodes can be
- //provided to the test.
- //Files are uploaded to nodes, other nodes try to retrieve the file
- //Number of nodes can be provided via commandline too.
- func TestFileRetrieval(t *testing.T) {
- if *nodes != 0 {
- err := runFileRetrievalTest(*nodes)
- if err != nil {
- t.Fatal(err)
- }
- } else {
- nodeCnt := []int{16}
- //if the `longrunning` flag has been provided
- //run more test combinations
- if *longrunning {
- nodeCnt = append(nodeCnt, 32, 64, 128)
- }
- for _, n := range nodeCnt {
- err := runFileRetrievalTest(n)
- if err != nil {
- t.Fatal(err)
- }
- }
- }
- }
- //This test is a retrieval test for nodes.
- //One node is randomly selected to be the pivot node.
- //A configurable number of chunks and nodes can be
- //provided to the test, the number of chunks is uploaded
- //to the pivot node and other nodes try to retrieve the chunk(s).
- //Number of chunks and nodes can be provided via commandline too.
- func TestRetrieval(t *testing.T) {
- //if nodes/chunks have been provided via commandline,
- //run the tests with these values
- if *nodes != 0 && *chunks != 0 {
- err := runRetrievalTest(*chunks, *nodes)
- if err != nil {
- t.Fatal(err)
- }
- } else {
- var nodeCnt []int
- var chnkCnt []int
- //if the `longrunning` flag has been provided
- //run more test combinations
- if *longrunning {
- nodeCnt = []int{16, 32, 128}
- chnkCnt = []int{4, 32, 256}
- } else {
- //default test
- nodeCnt = []int{16}
- chnkCnt = []int{32}
- }
- for _, n := range nodeCnt {
- for _, c := range chnkCnt {
- err := runRetrievalTest(c, n)
- if err != nil {
- t.Fatal(err)
- }
- }
- }
- }
- }
- var retrievalSimServiceMap = map[string]simulation.ServiceFunc{
- "streamer": retrievalStreamerFunc,
- }
- func retrievalStreamerFunc(ctx *adapters.ServiceContext, bucket *sync.Map) (s node.Service, cleanup func(), err error) {
- n := ctx.Config.Node()
- addr := network.NewAddr(n)
- store, datadir, err := createTestLocalStorageForID(n.ID(), addr)
- if err != nil {
- return nil, nil, err
- }
- bucket.Store(bucketKeyStore, store)
- localStore := store.(*storage.LocalStore)
- netStore, err := storage.NewNetStore(localStore, nil)
- if err != nil {
- return nil, nil, err
- }
- kad := network.NewKademlia(addr.Over(), network.NewKadParams())
- delivery := NewDelivery(kad, netStore)
- netStore.NewNetFetcherFunc = network.NewFetcherFactory(delivery.RequestFromPeers, true).New
- r := NewRegistry(addr.ID(), delivery, netStore, state.NewInmemoryStore(), &RegistryOptions{
- DoSync: true,
- SyncUpdateDelay: 3 * time.Second,
- DoRetrieve: true,
- })
- fileStore := storage.NewFileStore(netStore, storage.NewFileStoreParams())
- bucket.Store(bucketKeyFileStore, fileStore)
- cleanup = func() {
- os.RemoveAll(datadir)
- netStore.Close()
- r.Close()
- }
- return r, cleanup, nil
- }
- /*
- The test loads a snapshot file to construct the swarm network,
- assuming that the snapshot file identifies a healthy
- kademlia network. Nevertheless a health check runs in the
- simulation's `action` function.
- The snapshot should have 'streamer' in its service list.
- */
- func runFileRetrievalTest(nodeCount int) error {
- sim := simulation.New(retrievalSimServiceMap)
- defer sim.Close()
- log.Info("Initializing test config")
- conf := &synctestConfig{}
- //map of discover ID to indexes of chunks expected at that ID
- conf.idToChunksMap = make(map[enode.ID][]int)
- //map of overlay address to discover ID
- conf.addrToIDMap = make(map[string]enode.ID)
- //array where the generated chunk hashes will be stored
- conf.hashes = make([]storage.Address, 0)
- err := sim.UploadSnapshot(fmt.Sprintf("testing/snapshot_%d.json", nodeCount))
- if err != nil {
- return err
- }
- ctx, cancelSimRun := context.WithTimeout(context.Background(), 1*time.Minute)
- defer cancelSimRun()
- result := sim.Run(ctx, func(ctx context.Context, sim *simulation.Simulation) error {
- nodeIDs := sim.UpNodeIDs()
- for _, n := range nodeIDs {
- //get the kademlia overlay address from this ID
- a := n.Bytes()
- //append it to the array of all overlay addresses
- conf.addrs = append(conf.addrs, a)
- //the proximity calculation is on overlay addr,
- //the p2p/simulations check func triggers on enode.ID,
- //so we need to know which overlay addr maps to which nodeID
- conf.addrToIDMap[string(a)] = n
- }
- //an array for the random files
- var randomFiles []string
- //channel to signal when the upload has finished
- //uploadFinished := make(chan struct{})
- //channel to trigger new node checks
- conf.hashes, randomFiles, err = uploadFilesToNodes(sim)
- if err != nil {
- return err
- }
- if _, err := sim.WaitTillHealthy(ctx, 2); err != nil {
- return err
- }
- // File retrieval check is repeated until all uploaded files are retrieved from all nodes
- // or until the timeout is reached.
- REPEAT:
- for {
- for _, id := range nodeIDs {
- //for each expected file, check if it is in the local store
- item, ok := sim.NodeItem(id, bucketKeyFileStore)
- if !ok {
- return fmt.Errorf("No filestore")
- }
- fileStore := item.(*storage.FileStore)
- //check all chunks
- for i, hash := range conf.hashes {
- reader, _ := fileStore.Retrieve(context.TODO(), hash)
- //check that we can read the file size and that it corresponds to the generated file size
- if s, err := reader.Size(ctx, nil); err != nil || s != int64(len(randomFiles[i])) {
- log.Warn("Retrieve error", "err", err, "hash", hash, "nodeId", id)
- time.Sleep(500 * time.Millisecond)
- continue REPEAT
- }
- log.Debug(fmt.Sprintf("File with root hash %x successfully retrieved", hash))
- }
- }
- return nil
- }
- })
- if result.Error != nil {
- return result.Error
- }
- return nil
- }
- /*
- The test generates the given number of chunks.
- The test loads a snapshot file to construct the swarm network,
- assuming that the snapshot file identifies a healthy
- kademlia network. Nevertheless a health check runs in the
- simulation's `action` function.
- The snapshot should have 'streamer' in its service list.
- */
- func runRetrievalTest(chunkCount int, nodeCount int) error {
- sim := simulation.New(retrievalSimServiceMap)
- defer sim.Close()
- conf := &synctestConfig{}
- //map of discover ID to indexes of chunks expected at that ID
- conf.idToChunksMap = make(map[enode.ID][]int)
- //map of overlay address to discover ID
- conf.addrToIDMap = make(map[string]enode.ID)
- //array where the generated chunk hashes will be stored
- conf.hashes = make([]storage.Address, 0)
- err := sim.UploadSnapshot(fmt.Sprintf("testing/snapshot_%d.json", nodeCount))
- if err != nil {
- return err
- }
- ctx := context.Background()
- result := sim.Run(ctx, func(ctx context.Context, sim *simulation.Simulation) error {
- nodeIDs := sim.UpNodeIDs()
- for _, n := range nodeIDs {
- //get the kademlia overlay address from this ID
- a := n.Bytes()
- //append it to the array of all overlay addresses
- conf.addrs = append(conf.addrs, a)
- //the proximity calculation is on overlay addr,
- //the p2p/simulations check func triggers on enode.ID,
- //so we need to know which overlay addr maps to which nodeID
- conf.addrToIDMap[string(a)] = n
- }
- //this is the node selected for upload
- node := sim.RandomUpNode()
- item, ok := sim.NodeItem(node.ID, bucketKeyStore)
- if !ok {
- return fmt.Errorf("No localstore")
- }
- lstore := item.(*storage.LocalStore)
- conf.hashes, err = uploadFileToSingleNodeStore(node.ID, chunkCount, lstore)
- if err != nil {
- return err
- }
- if _, err := sim.WaitTillHealthy(ctx, 2); err != nil {
- return err
- }
- // File retrieval check is repeated until all uploaded files are retrieved from all nodes
- // or until the timeout is reached.
- REPEAT:
- for {
- for _, id := range nodeIDs {
- //for each expected chunk, check if it is in the local store
- //check on the node's FileStore (netstore)
- item, ok := sim.NodeItem(id, bucketKeyFileStore)
- if !ok {
- return fmt.Errorf("No filestore")
- }
- fileStore := item.(*storage.FileStore)
- //check all chunks
- for _, hash := range conf.hashes {
- reader, _ := fileStore.Retrieve(context.TODO(), hash)
- //check that we can read the chunk size and that it corresponds to the generated chunk size
- if s, err := reader.Size(ctx, nil); err != nil || s != int64(chunkSize) {
- log.Warn("Retrieve error", "err", err, "hash", hash, "nodeId", id, "size", s)
- time.Sleep(500 * time.Millisecond)
- continue REPEAT
- }
- log.Debug(fmt.Sprintf("Chunk with root hash %x successfully retrieved", hash))
- }
- }
- // all nodes and files found, exit loop and return without error
- return nil
- }
- })
- if result.Error != nil {
- return result.Error
- }
- return nil
- }
|