| 123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522 |
- // Copyright 2016 The go-ethereum Authors
- // This file is part of the go-ethereum library.
- //
- // The go-ethereum library is free software: you can redistribute it and/or modify
- // it under the terms of the GNU Lesser General Public License as published by
- // the Free Software Foundation, either version 3 of the License, or
- // (at your option) any later version.
- //
- // The go-ethereum library is distributed in the hope that it will be useful,
- // but WITHOUT ANY WARRANTY; without even the implied warranty of
- // MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
- // GNU Lesser General Public License for more details.
- //
- // You should have received a copy of the GNU Lesser General Public License
- // along with the go-ethereum library. If not, see <http://www.gnu.org/licenses/>.
- package storage
- import (
- "bytes"
- "fmt"
- "io/ioutil"
- "os"
- "sync"
- "testing"
- "time"
- "github.com/ethereum/go-ethereum/common"
- "github.com/ethereum/go-ethereum/swarm/log"
- "github.com/ethereum/go-ethereum/swarm/storage/mock/mem"
- ldberrors "github.com/syndtr/goleveldb/leveldb/errors"
- )
- type testDbStore struct {
- *LDBStore
- dir string
- }
- func newTestDbStore(mock bool, trusted bool) (*testDbStore, func(), error) {
- dir, err := ioutil.TempDir("", "bzz-storage-test")
- if err != nil {
- return nil, func() {}, err
- }
- var db *LDBStore
- storeparams := NewDefaultStoreParams()
- params := NewLDBStoreParams(storeparams, dir)
- params.Po = testPoFunc
- if mock {
- globalStore := mem.NewGlobalStore()
- addr := common.HexToAddress("0x5aaeb6053f3e94c9b9a09f33669435e7ef1beaed")
- mockStore := globalStore.NewNodeStore(addr)
- db, err = NewMockDbStore(params, mockStore)
- } else {
- db, err = NewLDBStore(params)
- }
- cleanup := func() {
- if db != nil {
- db.Close()
- }
- err = os.RemoveAll(dir)
- if err != nil {
- panic(fmt.Sprintf("db cleanup failed: %v", err))
- }
- }
- return &testDbStore{db, dir}, cleanup, err
- }
- func testPoFunc(k Address) (ret uint8) {
- basekey := make([]byte, 32)
- return uint8(Proximity(basekey[:], k[:]))
- }
- func (db *testDbStore) close() {
- db.Close()
- err := os.RemoveAll(db.dir)
- if err != nil {
- panic(err)
- }
- }
- func testDbStoreRandom(n int, processors int, chunksize int64, mock bool, t *testing.T) {
- db, cleanup, err := newTestDbStore(mock, true)
- defer cleanup()
- if err != nil {
- t.Fatalf("init dbStore failed: %v", err)
- }
- testStoreRandom(db, processors, n, chunksize, t)
- }
- func testDbStoreCorrect(n int, processors int, chunksize int64, mock bool, t *testing.T) {
- db, cleanup, err := newTestDbStore(mock, false)
- defer cleanup()
- if err != nil {
- t.Fatalf("init dbStore failed: %v", err)
- }
- testStoreCorrect(db, processors, n, chunksize, t)
- }
- func TestDbStoreRandom_1(t *testing.T) {
- testDbStoreRandom(1, 1, 0, false, t)
- }
- func TestDbStoreCorrect_1(t *testing.T) {
- testDbStoreCorrect(1, 1, 4096, false, t)
- }
- func TestDbStoreRandom_1_5k(t *testing.T) {
- testDbStoreRandom(8, 5000, 0, false, t)
- }
- func TestDbStoreRandom_8_5k(t *testing.T) {
- testDbStoreRandom(8, 5000, 0, false, t)
- }
- func TestDbStoreCorrect_1_5k(t *testing.T) {
- testDbStoreCorrect(1, 5000, 4096, false, t)
- }
- func TestDbStoreCorrect_8_5k(t *testing.T) {
- testDbStoreCorrect(8, 5000, 4096, false, t)
- }
- func TestMockDbStoreRandom_1(t *testing.T) {
- testDbStoreRandom(1, 1, 0, true, t)
- }
- func TestMockDbStoreCorrect_1(t *testing.T) {
- testDbStoreCorrect(1, 1, 4096, true, t)
- }
- func TestMockDbStoreRandom_1_5k(t *testing.T) {
- testDbStoreRandom(8, 5000, 0, true, t)
- }
- func TestMockDbStoreRandom_8_5k(t *testing.T) {
- testDbStoreRandom(8, 5000, 0, true, t)
- }
- func TestMockDbStoreCorrect_1_5k(t *testing.T) {
- testDbStoreCorrect(1, 5000, 4096, true, t)
- }
- func TestMockDbStoreCorrect_8_5k(t *testing.T) {
- testDbStoreCorrect(8, 5000, 4096, true, t)
- }
- func testDbStoreNotFound(t *testing.T, mock bool) {
- db, cleanup, err := newTestDbStore(mock, false)
- defer cleanup()
- if err != nil {
- t.Fatalf("init dbStore failed: %v", err)
- }
- _, err = db.Get(ZeroAddr)
- if err != ErrChunkNotFound {
- t.Errorf("Expected ErrChunkNotFound, got %v", err)
- }
- }
- func TestDbStoreNotFound(t *testing.T) {
- testDbStoreNotFound(t, false)
- }
- func TestMockDbStoreNotFound(t *testing.T) {
- testDbStoreNotFound(t, true)
- }
- func testIterator(t *testing.T, mock bool) {
- var chunkcount int = 32
- var i int
- var poc uint
- chunkkeys := NewAddressCollection(chunkcount)
- chunkkeys_results := NewAddressCollection(chunkcount)
- db, cleanup, err := newTestDbStore(mock, false)
- defer cleanup()
- if err != nil {
- t.Fatalf("init dbStore failed: %v", err)
- }
- chunks := GenerateRandomChunks(DefaultChunkSize, chunkcount)
- wg := &sync.WaitGroup{}
- wg.Add(len(chunks))
- for i = 0; i < len(chunks); i++ {
- db.Put(chunks[i])
- chunkkeys[i] = chunks[i].Addr
- j := i
- go func() {
- defer wg.Done()
- <-chunks[j].dbStoredC
- }()
- }
- //testSplit(m, l, 128, chunkkeys, t)
- for i = 0; i < len(chunkkeys); i++ {
- log.Trace(fmt.Sprintf("Chunk array pos %d/%d: '%v'", i, chunkcount, chunkkeys[i]))
- }
- wg.Wait()
- i = 0
- for poc = 0; poc <= 255; poc++ {
- err := db.SyncIterator(0, uint64(chunkkeys.Len()), uint8(poc), func(k Address, n uint64) bool {
- log.Trace(fmt.Sprintf("Got key %v number %d poc %d", k, n, uint8(poc)))
- chunkkeys_results[n-1] = k
- i++
- return true
- })
- if err != nil {
- t.Fatalf("Iterator call failed: %v", err)
- }
- }
- for i = 0; i < chunkcount; i++ {
- if !bytes.Equal(chunkkeys[i], chunkkeys_results[i]) {
- t.Fatalf("Chunk put #%d key '%v' does not match iterator's key '%v'", i, chunkkeys[i], chunkkeys_results[i])
- }
- }
- }
- func TestIterator(t *testing.T) {
- testIterator(t, false)
- }
- func TestMockIterator(t *testing.T) {
- testIterator(t, true)
- }
- func benchmarkDbStorePut(n int, processors int, chunksize int64, mock bool, b *testing.B) {
- db, cleanup, err := newTestDbStore(mock, true)
- defer cleanup()
- if err != nil {
- b.Fatalf("init dbStore failed: %v", err)
- }
- benchmarkStorePut(db, processors, n, chunksize, b)
- }
- func benchmarkDbStoreGet(n int, processors int, chunksize int64, mock bool, b *testing.B) {
- db, cleanup, err := newTestDbStore(mock, true)
- defer cleanup()
- if err != nil {
- b.Fatalf("init dbStore failed: %v", err)
- }
- benchmarkStoreGet(db, processors, n, chunksize, b)
- }
- func BenchmarkDbStorePut_1_500(b *testing.B) {
- benchmarkDbStorePut(500, 1, 4096, false, b)
- }
- func BenchmarkDbStorePut_8_500(b *testing.B) {
- benchmarkDbStorePut(500, 8, 4096, false, b)
- }
- func BenchmarkDbStoreGet_1_500(b *testing.B) {
- benchmarkDbStoreGet(500, 1, 4096, false, b)
- }
- func BenchmarkDbStoreGet_8_500(b *testing.B) {
- benchmarkDbStoreGet(500, 8, 4096, false, b)
- }
- func BenchmarkMockDbStorePut_1_500(b *testing.B) {
- benchmarkDbStorePut(500, 1, 4096, true, b)
- }
- func BenchmarkMockDbStorePut_8_500(b *testing.B) {
- benchmarkDbStorePut(500, 8, 4096, true, b)
- }
- func BenchmarkMockDbStoreGet_1_500(b *testing.B) {
- benchmarkDbStoreGet(500, 1, 4096, true, b)
- }
- func BenchmarkMockDbStoreGet_8_500(b *testing.B) {
- benchmarkDbStoreGet(500, 8, 4096, true, b)
- }
- // TestLDBStoreWithoutCollectGarbage tests that we can put a number of random chunks in the LevelDB store, and
- // retrieve them, provided we don't hit the garbage collection
- func TestLDBStoreWithoutCollectGarbage(t *testing.T) {
- capacity := 50
- n := 10
- ldb, cleanup := newLDBStore(t)
- ldb.setCapacity(uint64(capacity))
- defer cleanup()
- chunks := []*Chunk{}
- for i := 0; i < n; i++ {
- c := GenerateRandomChunk(DefaultChunkSize)
- chunks = append(chunks, c)
- log.Trace("generate random chunk", "idx", i, "chunk", c)
- }
- for i := 0; i < n; i++ {
- go ldb.Put(chunks[i])
- }
- // wait for all chunks to be stored
- for i := 0; i < n; i++ {
- <-chunks[i].dbStoredC
- }
- log.Info("ldbstore", "entrycnt", ldb.entryCnt, "accesscnt", ldb.accessCnt)
- for i := 0; i < n; i++ {
- ret, err := ldb.Get(chunks[i].Addr)
- if err != nil {
- t.Fatal(err)
- }
- if !bytes.Equal(ret.SData, chunks[i].SData) {
- t.Fatal("expected to get the same data back, but got smth else")
- }
- log.Info("got back chunk", "chunk", ret)
- }
- if ldb.entryCnt != uint64(n+1) {
- t.Fatalf("expected entryCnt to be equal to %v, but got %v", n+1, ldb.entryCnt)
- }
- if ldb.accessCnt != uint64(2*n+1) {
- t.Fatalf("expected accessCnt to be equal to %v, but got %v", n+1, ldb.accessCnt)
- }
- }
- // TestLDBStoreCollectGarbage tests that we can put more chunks than LevelDB's capacity, and
- // retrieve only some of them, because garbage collection must have cleared some of them
- func TestLDBStoreCollectGarbage(t *testing.T) {
- capacity := 500
- n := 2000
- ldb, cleanup := newLDBStore(t)
- ldb.setCapacity(uint64(capacity))
- defer cleanup()
- chunks := []*Chunk{}
- for i := 0; i < n; i++ {
- c := GenerateRandomChunk(DefaultChunkSize)
- chunks = append(chunks, c)
- log.Trace("generate random chunk", "idx", i, "chunk", c)
- }
- for i := 0; i < n; i++ {
- ldb.Put(chunks[i])
- }
- // wait for all chunks to be stored
- for i := 0; i < n; i++ {
- <-chunks[i].dbStoredC
- }
- log.Info("ldbstore", "entrycnt", ldb.entryCnt, "accesscnt", ldb.accessCnt)
- // wait for garbage collection to kick in on the responsible actor
- time.Sleep(5 * time.Second)
- var missing int
- for i := 0; i < n; i++ {
- ret, err := ldb.Get(chunks[i].Addr)
- if err == ErrChunkNotFound || err == ldberrors.ErrNotFound {
- missing++
- continue
- }
- if err != nil {
- t.Fatal(err)
- }
- if !bytes.Equal(ret.SData, chunks[i].SData) {
- t.Fatal("expected to get the same data back, but got smth else")
- }
- log.Trace("got back chunk", "chunk", ret)
- }
- if missing < n-capacity {
- t.Fatalf("gc failure: expected to miss %v chunks, but only %v are actually missing", n-capacity, missing)
- }
- log.Info("ldbstore", "total", n, "missing", missing, "entrycnt", ldb.entryCnt, "accesscnt", ldb.accessCnt)
- }
- // TestLDBStoreAddRemove tests that we can put and then delete a given chunk
- func TestLDBStoreAddRemove(t *testing.T) {
- ldb, cleanup := newLDBStore(t)
- ldb.setCapacity(200)
- defer cleanup()
- n := 100
- chunks := []*Chunk{}
- for i := 0; i < n; i++ {
- c := GenerateRandomChunk(DefaultChunkSize)
- chunks = append(chunks, c)
- log.Trace("generate random chunk", "idx", i, "chunk", c)
- }
- for i := 0; i < n; i++ {
- go ldb.Put(chunks[i])
- }
- // wait for all chunks to be stored before continuing
- for i := 0; i < n; i++ {
- <-chunks[i].dbStoredC
- }
- for i := 0; i < n; i++ {
- // delete all even index chunks
- if i%2 == 0 {
- key := chunks[i].Addr
- ikey := getIndexKey(key)
- var indx dpaDBIndex
- ldb.tryAccessIdx(ikey, &indx)
- ldb.delete(indx.Idx, ikey, ldb.po(key))
- }
- }
- log.Info("ldbstore", "entrycnt", ldb.entryCnt, "accesscnt", ldb.accessCnt)
- for i := 0; i < n; i++ {
- ret, err := ldb.Get(chunks[i].Addr)
- if i%2 == 0 {
- // expect even chunks to be missing
- if err == nil || ret != nil {
- t.Fatal("expected chunk to be missing, but got no error")
- }
- } else {
- // expect odd chunks to be retrieved successfully
- if err != nil {
- t.Fatalf("expected no error, but got %s", err)
- }
- if !bytes.Equal(ret.SData, chunks[i].SData) {
- t.Fatal("expected to get the same data back, but got smth else")
- }
- }
- }
- }
- // TestLDBStoreRemoveThenCollectGarbage tests that we can delete chunks and that we can trigger garbage collection
- func TestLDBStoreRemoveThenCollectGarbage(t *testing.T) {
- capacity := 10
- ldb, cleanup := newLDBStore(t)
- ldb.setCapacity(uint64(capacity))
- n := 7
- chunks := []*Chunk{}
- for i := 0; i < capacity; i++ {
- c := GenerateRandomChunk(DefaultChunkSize)
- chunks = append(chunks, c)
- log.Trace("generate random chunk", "idx", i, "chunk", c)
- }
- for i := 0; i < n; i++ {
- ldb.Put(chunks[i])
- }
- // wait for all chunks to be stored before continuing
- for i := 0; i < n; i++ {
- <-chunks[i].dbStoredC
- }
- // delete all chunks
- for i := 0; i < n; i++ {
- key := chunks[i].Addr
- ikey := getIndexKey(key)
- var indx dpaDBIndex
- ldb.tryAccessIdx(ikey, &indx)
- ldb.delete(indx.Idx, ikey, ldb.po(key))
- }
- log.Info("ldbstore", "entrycnt", ldb.entryCnt, "accesscnt", ldb.accessCnt)
- cleanup()
- ldb, cleanup = newLDBStore(t)
- ldb.setCapacity(uint64(capacity))
- n = 10
- for i := 0; i < n; i++ {
- ldb.Put(chunks[i])
- }
- // wait for all chunks to be stored before continuing
- for i := 0; i < n; i++ {
- <-chunks[i].dbStoredC
- }
- // expect for first chunk to be missing, because it has the smallest access value
- idx := 0
- ret, err := ldb.Get(chunks[idx].Addr)
- if err == nil || ret != nil {
- t.Fatal("expected first chunk to be missing, but got no error")
- }
- // expect for last chunk to be present, as it has the largest access value
- idx = 9
- ret, err = ldb.Get(chunks[idx].Addr)
- if err != nil {
- t.Fatalf("expected no error, but got %s", err)
- }
- if !bytes.Equal(ret.SData, chunks[idx].SData) {
- t.Fatal("expected to get the same data back, but got smth else")
- }
- }
|