| 123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743744745746747748749750751752753754755756757758759760761762763764765766767768769770771772773774775776777 |
- // Copyright 2016 The go-ethereum Authors
- // This file is part of the go-ethereum library.
- //
- // The go-ethereum library is free software: you can redistribute it and/or modify
- // it under the terms of the GNU Lesser General Public License as published by
- // the Free Software Foundation, either version 3 of the License, or
- // (at your option) any later version.
- //
- // The go-ethereum library is distributed in the hope that it will be useful,
- // but WITHOUT ANY WARRANTY; without even the implied warranty of
- // MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
- // GNU Lesser General Public License for more details.
- //
- // You should have received a copy of the GNU Lesser General Public License
- // along with the go-ethereum library. If not, see <http://www.gnu.org/licenses/>.
- package storage
- import (
- "bytes"
- "context"
- "encoding/binary"
- "fmt"
- "io/ioutil"
- "os"
- "strconv"
- "strings"
- "testing"
- "github.com/ethereum/go-ethereum/common"
- ch "github.com/ethereum/go-ethereum/swarm/chunk"
- "github.com/ethereum/go-ethereum/swarm/log"
- "github.com/ethereum/go-ethereum/swarm/storage/mock/mem"
- ldberrors "github.com/syndtr/goleveldb/leveldb/errors"
- )
- type testDbStore struct {
- *LDBStore
- dir string
- }
- func newTestDbStore(mock bool, trusted bool) (*testDbStore, func(), error) {
- dir, err := ioutil.TempDir("", "bzz-storage-test")
- if err != nil {
- return nil, func() {}, err
- }
- var db *LDBStore
- storeparams := NewDefaultStoreParams()
- params := NewLDBStoreParams(storeparams, dir)
- params.Po = testPoFunc
- if mock {
- globalStore := mem.NewGlobalStore()
- addr := common.HexToAddress("0x5aaeb6053f3e94c9b9a09f33669435e7ef1beaed")
- mockStore := globalStore.NewNodeStore(addr)
- db, err = NewMockDbStore(params, mockStore)
- } else {
- db, err = NewLDBStore(params)
- }
- cleanup := func() {
- if db != nil {
- db.Close()
- }
- err = os.RemoveAll(dir)
- if err != nil {
- panic(fmt.Sprintf("db cleanup failed: %v", err))
- }
- }
- return &testDbStore{db, dir}, cleanup, err
- }
- func testPoFunc(k Address) (ret uint8) {
- basekey := make([]byte, 32)
- return uint8(Proximity(basekey, k[:]))
- }
- func testDbStoreRandom(n int, mock bool, t *testing.T) {
- db, cleanup, err := newTestDbStore(mock, true)
- defer cleanup()
- if err != nil {
- t.Fatalf("init dbStore failed: %v", err)
- }
- testStoreRandom(db, n, t)
- }
- func testDbStoreCorrect(n int, mock bool, t *testing.T) {
- db, cleanup, err := newTestDbStore(mock, false)
- defer cleanup()
- if err != nil {
- t.Fatalf("init dbStore failed: %v", err)
- }
- testStoreCorrect(db, n, t)
- }
- func TestMarkAccessed(t *testing.T) {
- db, cleanup, err := newTestDbStore(false, true)
- defer cleanup()
- if err != nil {
- t.Fatalf("init dbStore failed: %v", err)
- }
- h := GenerateRandomChunk(ch.DefaultSize)
- db.Put(context.Background(), h)
- var index dpaDBIndex
- addr := h.Address()
- idxk := getIndexKey(addr)
- idata, err := db.db.Get(idxk)
- if err != nil {
- t.Fatal(err)
- }
- decodeIndex(idata, &index)
- if index.Access != 0 {
- t.Fatalf("Expected the access index to be %d, but it is %d", 0, index.Access)
- }
- db.MarkAccessed(addr)
- db.writeCurrentBatch()
- idata, err = db.db.Get(idxk)
- if err != nil {
- t.Fatal(err)
- }
- decodeIndex(idata, &index)
- if index.Access != 1 {
- t.Fatalf("Expected the access index to be %d, but it is %d", 1, index.Access)
- }
- }
- func TestDbStoreRandom_1(t *testing.T) {
- testDbStoreRandom(1, false, t)
- }
- func TestDbStoreCorrect_1(t *testing.T) {
- testDbStoreCorrect(1, false, t)
- }
- func TestDbStoreRandom_1k(t *testing.T) {
- testDbStoreRandom(1000, false, t)
- }
- func TestDbStoreCorrect_1k(t *testing.T) {
- testDbStoreCorrect(1000, false, t)
- }
- func TestMockDbStoreRandom_1(t *testing.T) {
- testDbStoreRandom(1, true, t)
- }
- func TestMockDbStoreCorrect_1(t *testing.T) {
- testDbStoreCorrect(1, true, t)
- }
- func TestMockDbStoreRandom_1k(t *testing.T) {
- testDbStoreRandom(1000, true, t)
- }
- func TestMockDbStoreCorrect_1k(t *testing.T) {
- testDbStoreCorrect(1000, true, t)
- }
- func testDbStoreNotFound(t *testing.T, mock bool) {
- db, cleanup, err := newTestDbStore(mock, false)
- defer cleanup()
- if err != nil {
- t.Fatalf("init dbStore failed: %v", err)
- }
- _, err = db.Get(context.TODO(), ZeroAddr)
- if err != ErrChunkNotFound {
- t.Errorf("Expected ErrChunkNotFound, got %v", err)
- }
- }
- func TestDbStoreNotFound(t *testing.T) {
- testDbStoreNotFound(t, false)
- }
- func TestMockDbStoreNotFound(t *testing.T) {
- testDbStoreNotFound(t, true)
- }
- func testIterator(t *testing.T, mock bool) {
- var chunkcount int = 32
- var i int
- var poc uint
- chunkkeys := NewAddressCollection(chunkcount)
- chunkkeysResults := NewAddressCollection(chunkcount)
- db, cleanup, err := newTestDbStore(mock, false)
- defer cleanup()
- if err != nil {
- t.Fatalf("init dbStore failed: %v", err)
- }
- chunks := GenerateRandomChunks(ch.DefaultSize, chunkcount)
- for i = 0; i < len(chunks); i++ {
- chunkkeys[i] = chunks[i].Address()
- err := db.Put(context.TODO(), chunks[i])
- if err != nil {
- t.Fatalf("dbStore.Put failed: %v", err)
- }
- }
- for i = 0; i < len(chunkkeys); i++ {
- log.Trace(fmt.Sprintf("Chunk array pos %d/%d: '%v'", i, chunkcount, chunkkeys[i]))
- }
- i = 0
- for poc = 0; poc <= 255; poc++ {
- err := db.SyncIterator(0, uint64(chunkkeys.Len()), uint8(poc), func(k Address, n uint64) bool {
- log.Trace(fmt.Sprintf("Got key %v number %d poc %d", k, n, uint8(poc)))
- chunkkeysResults[n] = k
- i++
- return true
- })
- if err != nil {
- t.Fatalf("Iterator call failed: %v", err)
- }
- }
- for i = 0; i < chunkcount; i++ {
- if !bytes.Equal(chunkkeys[i], chunkkeysResults[i]) {
- t.Fatalf("Chunk put #%d key '%v' does not match iterator's key '%v'", i, chunkkeys[i], chunkkeysResults[i])
- }
- }
- }
- func TestIterator(t *testing.T) {
- testIterator(t, false)
- }
- func TestMockIterator(t *testing.T) {
- testIterator(t, true)
- }
- func benchmarkDbStorePut(n int, mock bool, b *testing.B) {
- db, cleanup, err := newTestDbStore(mock, true)
- defer cleanup()
- if err != nil {
- b.Fatalf("init dbStore failed: %v", err)
- }
- benchmarkStorePut(db, n, b)
- }
- func benchmarkDbStoreGet(n int, mock bool, b *testing.B) {
- db, cleanup, err := newTestDbStore(mock, true)
- defer cleanup()
- if err != nil {
- b.Fatalf("init dbStore failed: %v", err)
- }
- benchmarkStoreGet(db, n, b)
- }
- func BenchmarkDbStorePut_500(b *testing.B) {
- benchmarkDbStorePut(500, false, b)
- }
- func BenchmarkDbStoreGet_500(b *testing.B) {
- benchmarkDbStoreGet(500, false, b)
- }
- func BenchmarkMockDbStorePut_500(b *testing.B) {
- benchmarkDbStorePut(500, true, b)
- }
- func BenchmarkMockDbStoreGet_500(b *testing.B) {
- benchmarkDbStoreGet(500, true, b)
- }
- // TestLDBStoreWithoutCollectGarbage tests that we can put a number of random chunks in the LevelDB store, and
- // retrieve them, provided we don't hit the garbage collection
- func TestLDBStoreWithoutCollectGarbage(t *testing.T) {
- capacity := 50
- n := 10
- ldb, cleanup := newLDBStore(t)
- ldb.setCapacity(uint64(capacity))
- defer cleanup()
- chunks, err := mputRandomChunks(ldb, n)
- if err != nil {
- t.Fatal(err.Error())
- }
- log.Info("ldbstore", "entrycnt", ldb.entryCnt, "accesscnt", ldb.accessCnt)
- for _, ch := range chunks {
- ret, err := ldb.Get(context.TODO(), ch.Address())
- if err != nil {
- t.Fatal(err)
- }
- if !bytes.Equal(ret.Data(), ch.Data()) {
- t.Fatal("expected to get the same data back, but got smth else")
- }
- }
- if ldb.entryCnt != uint64(n) {
- t.Fatalf("expected entryCnt to be equal to %v, but got %v", n, ldb.entryCnt)
- }
- if ldb.accessCnt != uint64(2*n) {
- t.Fatalf("expected accessCnt to be equal to %v, but got %v", 2*n, ldb.accessCnt)
- }
- }
- // TestLDBStoreCollectGarbage tests that we can put more chunks than LevelDB's capacity, and
- // retrieve only some of them, because garbage collection must have partially cleared the store
- // Also tests that we can delete chunks and that we can trigger garbage collection
- func TestLDBStoreCollectGarbage(t *testing.T) {
- // below max ronud
- initialCap := defaultMaxGCRound / 100
- cap := initialCap / 2
- t.Run(fmt.Sprintf("A/%d/%d", cap, cap*4), testLDBStoreCollectGarbage)
- t.Run(fmt.Sprintf("B/%d/%d", cap, cap*4), testLDBStoreRemoveThenCollectGarbage)
- // at max round
- cap = initialCap
- t.Run(fmt.Sprintf("A/%d/%d", cap, cap*4), testLDBStoreCollectGarbage)
- t.Run(fmt.Sprintf("B/%d/%d", cap, cap*4), testLDBStoreRemoveThenCollectGarbage)
- // more than max around, not on threshold
- cap = initialCap + 500
- t.Run(fmt.Sprintf("A/%d/%d", cap, cap*4), testLDBStoreCollectGarbage)
- t.Run(fmt.Sprintf("B/%d/%d", cap, cap*4), testLDBStoreRemoveThenCollectGarbage)
- }
- func testLDBStoreCollectGarbage(t *testing.T) {
- params := strings.Split(t.Name(), "/")
- capacity, err := strconv.Atoi(params[2])
- if err != nil {
- t.Fatal(err)
- }
- n, err := strconv.Atoi(params[3])
- if err != nil {
- t.Fatal(err)
- }
- ldb, cleanup := newLDBStore(t)
- ldb.setCapacity(uint64(capacity))
- defer cleanup()
- // retrieve the gc round target count for the db capacity
- ldb.startGC(capacity)
- roundTarget := ldb.gc.target
- // split put counts to gc target count threshold, and wait for gc to finish in between
- var allChunks []Chunk
- remaining := n
- for remaining > 0 {
- var putCount int
- if remaining < roundTarget {
- putCount = remaining
- } else {
- putCount = roundTarget
- }
- remaining -= putCount
- chunks, err := mputRandomChunks(ldb, putCount)
- if err != nil {
- t.Fatal(err.Error())
- }
- allChunks = append(allChunks, chunks...)
- ldb.lock.RLock()
- log.Debug("ldbstore", "entrycnt", ldb.entryCnt, "accesscnt", ldb.accessCnt, "cap", capacity, "n", n)
- ldb.lock.RUnlock()
- waitGc(ldb)
- }
- // attempt gets on all put chunks
- var missing int
- for _, ch := range allChunks {
- ret, err := ldb.Get(context.TODO(), ch.Address())
- if err == ErrChunkNotFound || err == ldberrors.ErrNotFound {
- missing++
- continue
- }
- if err != nil {
- t.Fatal(err)
- }
- if !bytes.Equal(ret.Data(), ch.Data()) {
- t.Fatal("expected to get the same data back, but got smth else")
- }
- log.Trace("got back chunk", "chunk", ret)
- }
- // all surplus chunks should be missing
- expectMissing := roundTarget + (((n - capacity) / roundTarget) * roundTarget)
- if missing != expectMissing {
- t.Fatalf("gc failure: expected to miss %v chunks, but only %v are actually missing", expectMissing, missing)
- }
- log.Info("ldbstore", "total", n, "missing", missing, "entrycnt", ldb.entryCnt, "accesscnt", ldb.accessCnt)
- }
- // TestLDBStoreAddRemove tests that we can put and then delete a given chunk
- func TestLDBStoreAddRemove(t *testing.T) {
- ldb, cleanup := newLDBStore(t)
- ldb.setCapacity(200)
- defer cleanup()
- n := 100
- chunks, err := mputRandomChunks(ldb, n)
- if err != nil {
- t.Fatalf(err.Error())
- }
- for i := 0; i < n; i++ {
- // delete all even index chunks
- if i%2 == 0 {
- ldb.Delete(chunks[i].Address())
- }
- }
- log.Info("ldbstore", "entrycnt", ldb.entryCnt, "accesscnt", ldb.accessCnt)
- for i := 0; i < n; i++ {
- ret, err := ldb.Get(context.TODO(), chunks[i].Address())
- if i%2 == 0 {
- // expect even chunks to be missing
- if err == nil {
- t.Fatal("expected chunk to be missing, but got no error")
- }
- } else {
- // expect odd chunks to be retrieved successfully
- if err != nil {
- t.Fatalf("expected no error, but got %s", err)
- }
- if !bytes.Equal(ret.Data(), chunks[i].Data()) {
- t.Fatal("expected to get the same data back, but got smth else")
- }
- }
- }
- }
- func testLDBStoreRemoveThenCollectGarbage(t *testing.T) {
- t.Skip("flaky with -race flag")
- params := strings.Split(t.Name(), "/")
- capacity, err := strconv.Atoi(params[2])
- if err != nil {
- t.Fatal(err)
- }
- n, err := strconv.Atoi(params[3])
- if err != nil {
- t.Fatal(err)
- }
- ldb, cleanup := newLDBStore(t)
- defer cleanup()
- ldb.setCapacity(uint64(capacity))
- // put capacity count number of chunks
- chunks := make([]Chunk, n)
- for i := 0; i < n; i++ {
- c := GenerateRandomChunk(ch.DefaultSize)
- chunks[i] = c
- log.Trace("generate random chunk", "idx", i, "chunk", c)
- }
- for i := 0; i < n; i++ {
- err := ldb.Put(context.TODO(), chunks[i])
- if err != nil {
- t.Fatal(err)
- }
- }
- waitGc(ldb)
- // delete all chunks
- // (only count the ones actually deleted, the rest will have been gc'd)
- deletes := 0
- for i := 0; i < n; i++ {
- if ldb.Delete(chunks[i].Address()) == nil {
- deletes++
- }
- }
- log.Info("ldbstore", "entrycnt", ldb.entryCnt, "accesscnt", ldb.accessCnt)
- if ldb.entryCnt != 0 {
- t.Fatalf("ldb.entrCnt expected 0 got %v", ldb.entryCnt)
- }
- // the manual deletes will have increased accesscnt, so we need to add this when we verify the current count
- expAccessCnt := uint64(n)
- if ldb.accessCnt != expAccessCnt {
- t.Fatalf("ldb.accessCnt expected %v got %v", expAccessCnt, ldb.accessCnt)
- }
- // retrieve the gc round target count for the db capacity
- ldb.startGC(capacity)
- roundTarget := ldb.gc.target
- remaining := n
- var puts int
- for remaining > 0 {
- var putCount int
- if remaining < roundTarget {
- putCount = remaining
- } else {
- putCount = roundTarget
- }
- remaining -= putCount
- for putCount > 0 {
- ldb.Put(context.TODO(), chunks[puts])
- ldb.lock.RLock()
- log.Debug("ldbstore", "entrycnt", ldb.entryCnt, "accesscnt", ldb.accessCnt, "cap", capacity, "n", n, "puts", puts, "remaining", remaining, "roundtarget", roundTarget)
- ldb.lock.RUnlock()
- puts++
- putCount--
- }
- waitGc(ldb)
- }
- // expect first surplus chunks to be missing, because they have the smallest access value
- expectMissing := roundTarget + (((n - capacity) / roundTarget) * roundTarget)
- for i := 0; i < expectMissing; i++ {
- _, err := ldb.Get(context.TODO(), chunks[i].Address())
- if err == nil {
- t.Fatalf("expected surplus chunk %d to be missing, but got no error", i)
- }
- }
- // expect last chunks to be present, as they have the largest access value
- for i := expectMissing; i < n; i++ {
- ret, err := ldb.Get(context.TODO(), chunks[i].Address())
- if err != nil {
- t.Fatalf("chunk %v: expected no error, but got %s", i, err)
- }
- if !bytes.Equal(ret.Data(), chunks[i].Data()) {
- t.Fatal("expected to get the same data back, but got smth else")
- }
- }
- }
- // TestLDBStoreCollectGarbageAccessUnlikeIndex tests garbage collection where accesscount differs from indexcount
- func TestLDBStoreCollectGarbageAccessUnlikeIndex(t *testing.T) {
- capacity := defaultMaxGCRound / 100 * 2
- n := capacity - 1
- ldb, cleanup := newLDBStore(t)
- ldb.setCapacity(uint64(capacity))
- defer cleanup()
- chunks, err := mputRandomChunks(ldb, n)
- if err != nil {
- t.Fatal(err.Error())
- }
- log.Info("ldbstore", "entrycnt", ldb.entryCnt, "accesscnt", ldb.accessCnt)
- // set first added capacity/2 chunks to highest accesscount
- for i := 0; i < capacity/2; i++ {
- _, err := ldb.Get(context.TODO(), chunks[i].Address())
- if err != nil {
- t.Fatalf("fail add chunk #%d - %s: %v", i, chunks[i].Address(), err)
- }
- }
- _, err = mputRandomChunks(ldb, 2)
- if err != nil {
- t.Fatal(err.Error())
- }
- // wait for garbage collection to kick in on the responsible actor
- waitGc(ldb)
- var missing int
- for i, ch := range chunks[2 : capacity/2] {
- ret, err := ldb.Get(context.TODO(), ch.Address())
- if err == ErrChunkNotFound || err == ldberrors.ErrNotFound {
- t.Fatalf("fail find chunk #%d - %s: %v", i, ch.Address(), err)
- }
- if !bytes.Equal(ret.Data(), ch.Data()) {
- t.Fatal("expected to get the same data back, but got smth else")
- }
- log.Trace("got back chunk", "chunk", ret)
- }
- log.Info("ldbstore", "total", n, "missing", missing, "entrycnt", ldb.entryCnt, "accesscnt", ldb.accessCnt)
- }
- func TestCleanIndex(t *testing.T) {
- capacity := 5000
- n := 3
- ldb, cleanup := newLDBStore(t)
- ldb.setCapacity(uint64(capacity))
- defer cleanup()
- chunks, err := mputRandomChunks(ldb, n)
- if err != nil {
- t.Fatal(err)
- }
- // remove the data of the first chunk
- po := ldb.po(chunks[0].Address()[:])
- dataKey := make([]byte, 10)
- dataKey[0] = keyData
- dataKey[1] = byte(po)
- // dataKey[2:10] = first chunk has storageIdx 0 on [2:10]
- if _, err := ldb.db.Get(dataKey); err != nil {
- t.Fatal(err)
- }
- if err := ldb.db.Delete(dataKey); err != nil {
- t.Fatal(err)
- }
- // remove the gc index row for the first chunk
- gcFirstCorrectKey := make([]byte, 9)
- gcFirstCorrectKey[0] = keyGCIdx
- if err := ldb.db.Delete(gcFirstCorrectKey); err != nil {
- t.Fatal(err)
- }
- // warp the gc data of the second chunk
- // this data should be correct again after the clean
- gcSecondCorrectKey := make([]byte, 9)
- gcSecondCorrectKey[0] = keyGCIdx
- binary.BigEndian.PutUint64(gcSecondCorrectKey[1:], uint64(1))
- gcSecondCorrectVal, err := ldb.db.Get(gcSecondCorrectKey)
- if err != nil {
- t.Fatal(err)
- }
- warpedGCVal := make([]byte, len(gcSecondCorrectVal)+1)
- copy(warpedGCVal[1:], gcSecondCorrectVal)
- if err := ldb.db.Delete(gcSecondCorrectKey); err != nil {
- t.Fatal(err)
- }
- if err := ldb.db.Put(gcSecondCorrectKey, warpedGCVal); err != nil {
- t.Fatal(err)
- }
- if err := ldb.CleanGCIndex(); err != nil {
- t.Fatal(err)
- }
- // the index without corresponding data should have been deleted
- idxKey := make([]byte, 33)
- idxKey[0] = keyIndex
- copy(idxKey[1:], chunks[0].Address())
- if _, err := ldb.db.Get(idxKey); err == nil {
- t.Fatalf("expected chunk 0 idx to be pruned: %v", idxKey)
- }
- // the two other indices should be present
- copy(idxKey[1:], chunks[1].Address())
- if _, err := ldb.db.Get(idxKey); err != nil {
- t.Fatalf("expected chunk 1 idx to be present: %v", idxKey)
- }
- copy(idxKey[1:], chunks[2].Address())
- if _, err := ldb.db.Get(idxKey); err != nil {
- t.Fatalf("expected chunk 2 idx to be present: %v", idxKey)
- }
- // first gc index should still be gone
- if _, err := ldb.db.Get(gcFirstCorrectKey); err == nil {
- t.Fatalf("expected gc 0 idx to be pruned: %v", idxKey)
- }
- // second gc index should still be fixed
- if _, err := ldb.db.Get(gcSecondCorrectKey); err != nil {
- t.Fatalf("expected gc 1 idx to be present: %v", idxKey)
- }
- // third gc index should be unchanged
- binary.BigEndian.PutUint64(gcSecondCorrectKey[1:], uint64(2))
- if _, err := ldb.db.Get(gcSecondCorrectKey); err != nil {
- t.Fatalf("expected gc 2 idx to be present: %v", idxKey)
- }
- c, err := ldb.db.Get(keyEntryCnt)
- if err != nil {
- t.Fatalf("expected gc 2 idx to be present: %v", idxKey)
- }
- // entrycount should now be one less
- entryCount := binary.BigEndian.Uint64(c)
- if entryCount != 2 {
- t.Fatalf("expected entrycnt to be 2, was %d", c)
- }
- // the chunks might accidentally be in the same bin
- // if so that bin counter will now be 2 - the highest added index.
- // if not, the total of them will be 3
- poBins := []uint8{ldb.po(chunks[1].Address()), ldb.po(chunks[2].Address())}
- if poBins[0] == poBins[1] {
- poBins = poBins[:1]
- }
- var binTotal uint64
- var currentBin [2]byte
- currentBin[0] = keyDistanceCnt
- if len(poBins) == 1 {
- currentBin[1] = poBins[0]
- c, err := ldb.db.Get(currentBin[:])
- if err != nil {
- t.Fatalf("expected gc 2 idx to be present: %v", idxKey)
- }
- binCount := binary.BigEndian.Uint64(c)
- if binCount != 2 {
- t.Fatalf("expected entrycnt to be 2, was %d", binCount)
- }
- } else {
- for _, bin := range poBins {
- currentBin[1] = bin
- c, err := ldb.db.Get(currentBin[:])
- if err != nil {
- t.Fatalf("expected gc 2 idx to be present: %v", idxKey)
- }
- binCount := binary.BigEndian.Uint64(c)
- binTotal += binCount
- }
- if binTotal != 3 {
- t.Fatalf("expected sum of bin indices to be 3, was %d", binTotal)
- }
- }
- // check that the iterator quits properly
- chunks, err = mputRandomChunks(ldb, 4100)
- if err != nil {
- t.Fatal(err)
- }
- po = ldb.po(chunks[4099].Address()[:])
- dataKey = make([]byte, 10)
- dataKey[0] = keyData
- dataKey[1] = byte(po)
- binary.BigEndian.PutUint64(dataKey[2:], 4099+3)
- if _, err := ldb.db.Get(dataKey); err != nil {
- t.Fatal(err)
- }
- if err := ldb.db.Delete(dataKey); err != nil {
- t.Fatal(err)
- }
- if err := ldb.CleanGCIndex(); err != nil {
- t.Fatal(err)
- }
- // entrycount should now be one less of added chunks
- c, err = ldb.db.Get(keyEntryCnt)
- if err != nil {
- t.Fatalf("expected gc 2 idx to be present: %v", idxKey)
- }
- entryCount = binary.BigEndian.Uint64(c)
- if entryCount != 4099+2 {
- t.Fatalf("expected entrycnt to be 2, was %d", c)
- }
- }
- // Note: waitGc does not guarantee that we wait 1 GC round; it only
- // guarantees that if the GC is running we wait for that run to finish
- // ticket: https://github.com/ethersphere/go-ethereum/issues/1151
- func waitGc(ldb *LDBStore) {
- <-ldb.gc.runC
- ldb.gc.runC <- struct{}{}
- }
|