| 123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448 |
- // Copyright 2019 The go-ethereum Authors
- // This file is part of the go-ethereum library.
- //
- // The go-ethereum library is free software: you can redistribute it and/or modify
- // it under the terms of the GNU Lesser General Public License as published by
- // the Free Software Foundation, either version 3 of the License, or
- // (at your option) any later version.
- //
- // The go-ethereum library is distributed in the hope that it will be useful,
- // but WITHOUT ANY WARRANTY; without even the implied warranty of
- // MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
- // GNU Lesser General Public License for more details.
- //
- // You should have received a copy of the GNU Lesser General Public License
- // along with the go-ethereum library. If not, see <http://www.gnu.org/licenses/>.
- package snapshot
- import (
- "bytes"
- "fmt"
- "math/big"
- "math/rand"
- "testing"
- "time"
- "github.com/allegro/bigcache"
- "github.com/ethereum/go-ethereum/common"
- "github.com/ethereum/go-ethereum/core/rawdb"
- "github.com/ethereum/go-ethereum/rlp"
- )
- func randomAccount() []byte {
- root := randomHash()
- a := Account{
- Balance: big.NewInt(rand.Int63()),
- Nonce: rand.Uint64(),
- Root: root[:],
- CodeHash: emptyCode[:],
- }
- data, _ := rlp.EncodeToBytes(a)
- return data
- }
- // TestMergeBasics tests some simple merges
- func TestMergeBasics(t *testing.T) {
- var (
- accounts = make(map[common.Hash][]byte)
- storage = make(map[common.Hash]map[common.Hash][]byte)
- )
- // Fill up a parent
- for i := 0; i < 100; i++ {
- h := randomHash()
- data := randomAccount()
- accounts[h] = data
- if rand.Intn(20) < 10 {
- accStorage := make(map[common.Hash][]byte)
- value := make([]byte, 32)
- rand.Read(value)
- accStorage[randomHash()] = value
- storage[h] = accStorage
- }
- }
- // Add some (identical) layers on top
- parent := newDiffLayer(emptyLayer{}, 1, common.Hash{}, accounts, storage)
- child := newDiffLayer(parent, 1, common.Hash{}, accounts, storage)
- child = newDiffLayer(child, 1, common.Hash{}, accounts, storage)
- child = newDiffLayer(child, 1, common.Hash{}, accounts, storage)
- child = newDiffLayer(child, 1, common.Hash{}, accounts, storage)
- // And flatten
- merged := (child.flatten()).(*diffLayer)
- { // Check account lists
- // Should be zero/nil first
- if got, exp := len(merged.accountList), 0; got != exp {
- t.Errorf("accountList wrong, got %v exp %v", got, exp)
- }
- // Then set when we call AccountList
- if got, exp := len(merged.AccountList()), len(accounts); got != exp {
- t.Errorf("AccountList() wrong, got %v exp %v", got, exp)
- }
- if got, exp := len(merged.accountList), len(accounts); got != exp {
- t.Errorf("accountList [2] wrong, got %v exp %v", got, exp)
- }
- }
- { // Check storage lists
- i := 0
- for aHash, sMap := range storage {
- if got, exp := len(merged.storageList), i; got != exp {
- t.Errorf("[1] storageList wrong, got %v exp %v", got, exp)
- }
- if got, exp := len(merged.StorageList(aHash)), len(sMap); got != exp {
- t.Errorf("[2] StorageList() wrong, got %v exp %v", got, exp)
- }
- if got, exp := len(merged.storageList[aHash]), len(sMap); got != exp {
- t.Errorf("storageList wrong, got %v exp %v", got, exp)
- }
- i++
- }
- }
- }
- // TestMergeDelete tests some deletion
- func TestMergeDelete(t *testing.T) {
- var (
- storage = make(map[common.Hash]map[common.Hash][]byte)
- )
- // Fill up a parent
- h1 := common.HexToHash("0x01")
- h2 := common.HexToHash("0x02")
- flip := func() map[common.Hash][]byte {
- accs := make(map[common.Hash][]byte)
- accs[h1] = randomAccount()
- accs[h2] = nil
- return accs
- }
- flop := func() map[common.Hash][]byte {
- accs := make(map[common.Hash][]byte)
- accs[h1] = nil
- accs[h2] = randomAccount()
- return accs
- }
- // Add some flip-flopping layers on top
- parent := newDiffLayer(emptyLayer{}, 1, common.Hash{}, flip(), storage)
- child := parent.Update(common.Hash{}, flop(), storage)
- child = child.Update(common.Hash{}, flip(), storage)
- child = child.Update(common.Hash{}, flop(), storage)
- child = child.Update(common.Hash{}, flip(), storage)
- child = child.Update(common.Hash{}, flop(), storage)
- child = child.Update(common.Hash{}, flip(), storage)
- if data, _ := child.Account(h1); data == nil {
- t.Errorf("last diff layer: expected %x to be non-nil", h1)
- }
- if data, _ := child.Account(h2); data != nil {
- t.Errorf("last diff layer: expected %x to be nil", h2)
- }
- // And flatten
- merged := (child.flatten()).(*diffLayer)
- // check number
- if got, exp := merged.number, child.number; got != exp {
- t.Errorf("merged layer: wrong number - exp %d got %d", exp, got)
- }
- if data, _ := merged.Account(h1); data == nil {
- t.Errorf("merged layer: expected %x to be non-nil", h1)
- }
- if data, _ := merged.Account(h2); data != nil {
- t.Errorf("merged layer: expected %x to be nil", h2)
- }
- // If we add more granular metering of memory, we can enable this again,
- // but it's not implemented for now
- //if got, exp := merged.memory, child.memory; got != exp {
- // t.Errorf("mem wrong, got %d, exp %d", got, exp)
- //}
- }
- // This tests that if we create a new account, and set a slot, and then merge
- // it, the lists will be correct.
- func TestInsertAndMerge(t *testing.T) {
- // Fill up a parent
- var (
- acc = common.HexToHash("0x01")
- slot = common.HexToHash("0x02")
- parent *diffLayer
- child *diffLayer
- )
- {
- var accounts = make(map[common.Hash][]byte)
- var storage = make(map[common.Hash]map[common.Hash][]byte)
- parent = newDiffLayer(emptyLayer{}, 1, common.Hash{}, accounts, storage)
- }
- {
- var accounts = make(map[common.Hash][]byte)
- var storage = make(map[common.Hash]map[common.Hash][]byte)
- accounts[acc] = randomAccount()
- accstorage := make(map[common.Hash][]byte)
- storage[acc] = accstorage
- storage[acc][slot] = []byte{0x01}
- child = newDiffLayer(parent, 2, common.Hash{}, accounts, storage)
- }
- // And flatten
- merged := (child.flatten()).(*diffLayer)
- { // Check that slot value is present
- got, _ := merged.Storage(acc, slot)
- if exp := []byte{0x01}; bytes.Compare(got, exp) != 0 {
- t.Errorf("merged slot value wrong, got %x, exp %x", got, exp)
- }
- }
- }
- // TestCapTree tests some functionality regarding capping/flattening
- func TestCapTree(t *testing.T) {
- var (
- storage = make(map[common.Hash]map[common.Hash][]byte)
- )
- setAccount := func(accKey string) map[common.Hash][]byte {
- return map[common.Hash][]byte{
- common.HexToHash(accKey): randomAccount(),
- }
- }
- // the bottom-most layer, aside from the 'disk layer'
- cache, _ := bigcache.NewBigCache(bigcache.Config{ // TODO(karalabe): dedup
- Shards: 1,
- LifeWindow: time.Hour,
- MaxEntriesInWindow: 1 * 1024,
- MaxEntrySize: 1,
- HardMaxCacheSize: 1,
- })
- base := &diskLayer{
- journal: "",
- db: rawdb.NewMemoryDatabase(),
- cache: cache,
- number: 0,
- root: common.HexToHash("0x01"),
- }
- // The lowest difflayer
- a1 := base.Update(common.HexToHash("0xa1"), setAccount("0xa1"), storage)
- a2 := a1.Update(common.HexToHash("0xa2"), setAccount("0xa2"), storage)
- b2 := a1.Update(common.HexToHash("0xb2"), setAccount("0xb2"), storage)
- a3 := a2.Update(common.HexToHash("0xa3"), setAccount("0xa3"), storage)
- b3 := b2.Update(common.HexToHash("0xb3"), setAccount("0xb3"), storage)
- checkExist := func(layer *diffLayer, key string) error {
- accountKey := common.HexToHash(key)
- data, _ := layer.Account(accountKey)
- if data == nil {
- return fmt.Errorf("expected %x to exist, got nil", accountKey)
- }
- return nil
- }
- shouldErr := func(layer *diffLayer, key string) error {
- accountKey := common.HexToHash(key)
- data, err := layer.Account(accountKey)
- if err == nil {
- return fmt.Errorf("expected error, got data %x", data)
- }
- return nil
- }
- // check basics
- if err := checkExist(b3, "0xa1"); err != nil {
- t.Error(err)
- }
- if err := checkExist(b3, "0xb2"); err != nil {
- t.Error(err)
- }
- if err := checkExist(b3, "0xb3"); err != nil {
- t.Error(err)
- }
- // Now, merge the a-chain
- diskNum, diffNum := a3.Cap(0, 1024)
- if diskNum != 0 {
- t.Errorf("disk layer err, got %d exp %d", diskNum, 0)
- }
- if diffNum != 2 {
- t.Errorf("diff layer err, got %d exp %d", diffNum, 2)
- }
- // At this point, a2 got merged into a1. Thus, a1 is now modified,
- // and as a1 is the parent of b2, b2 should no longer be able to iterate into parent
- // These should still be accessible
- if err := checkExist(b3, "0xb2"); err != nil {
- t.Error(err)
- }
- if err := checkExist(b3, "0xb3"); err != nil {
- t.Error(err)
- }
- //b2ParentNum, _ := b2.parent.Info()
- //if b2.parent.invalid == false
- // t.Errorf("err, exp parent to be invalid, got %v", b2.parent, b2ParentNum)
- //}
- // But these would need iteration into the modified parent:
- if err := shouldErr(b3, "0xa1"); err != nil {
- t.Error(err)
- }
- if err := shouldErr(b3, "0xa2"); err != nil {
- t.Error(err)
- }
- if err := shouldErr(b3, "0xa3"); err != nil {
- t.Error(err)
- }
- }
- type emptyLayer struct{}
- func (emptyLayer) Update(blockRoot common.Hash, accounts map[common.Hash][]byte, storage map[common.Hash]map[common.Hash][]byte) *diffLayer {
- panic("implement me")
- }
- func (emptyLayer) Cap(layers int, memory uint64) (uint64, uint64) {
- panic("implement me")
- }
- func (emptyLayer) Journal() error {
- panic("implement me")
- }
- func (emptyLayer) Info() (uint64, common.Hash) {
- return 0, common.Hash{}
- }
- func (emptyLayer) Number() uint64 {
- return 0
- }
- func (emptyLayer) Account(hash common.Hash) (*Account, error) {
- return nil, nil
- }
- func (emptyLayer) AccountRLP(hash common.Hash) ([]byte, error) {
- return nil, nil
- }
- func (emptyLayer) Storage(accountHash, storageHash common.Hash) ([]byte, error) {
- return nil, nil
- }
- // BenchmarkSearch checks how long it takes to find a non-existing key
- // BenchmarkSearch-6 200000 10481 ns/op (1K per layer)
- // BenchmarkSearch-6 200000 10760 ns/op (10K per layer)
- // BenchmarkSearch-6 100000 17866 ns/op
- //
- // BenchmarkSearch-6 500000 3723 ns/op (10k per layer, only top-level RLock()
- func BenchmarkSearch(b *testing.B) {
- // First, we set up 128 diff layers, with 1K items each
- blocknum := uint64(0)
- fill := func(parent snapshot) *diffLayer {
- accounts := make(map[common.Hash][]byte)
- storage := make(map[common.Hash]map[common.Hash][]byte)
- for i := 0; i < 10000; i++ {
- accounts[randomHash()] = randomAccount()
- }
- blocknum++
- return newDiffLayer(parent, blocknum, common.Hash{}, accounts, storage)
- }
- var layer snapshot
- layer = emptyLayer{}
- for i := 0; i < 128; i++ {
- layer = fill(layer)
- }
- key := common.Hash{}
- b.ResetTimer()
- for i := 0; i < b.N; i++ {
- layer.AccountRLP(key)
- }
- }
- // BenchmarkSearchSlot checks how long it takes to find a non-existing key
- // - Number of layers: 128
- // - Each layers contains the account, with a couple of storage slots
- // BenchmarkSearchSlot-6 100000 14554 ns/op
- // BenchmarkSearchSlot-6 100000 22254 ns/op (when checking parent root using mutex)
- // BenchmarkSearchSlot-6 100000 14551 ns/op (when checking parent number using atomic)
- func BenchmarkSearchSlot(b *testing.B) {
- // First, we set up 128 diff layers, with 1K items each
- blocknum := uint64(0)
- accountKey := common.Hash{}
- storageKey := common.HexToHash("0x1337")
- accountRLP := randomAccount()
- fill := func(parent snapshot) *diffLayer {
- accounts := make(map[common.Hash][]byte)
- accounts[accountKey] = accountRLP
- storage := make(map[common.Hash]map[common.Hash][]byte)
- accStorage := make(map[common.Hash][]byte)
- for i := 0; i < 5; i++ {
- value := make([]byte, 32)
- rand.Read(value)
- accStorage[randomHash()] = value
- storage[accountKey] = accStorage
- }
- blocknum++
- return newDiffLayer(parent, blocknum, common.Hash{}, accounts, storage)
- }
- var layer snapshot
- layer = emptyLayer{}
- for i := 0; i < 128; i++ {
- layer = fill(layer)
- }
- b.ResetTimer()
- for i := 0; i < b.N; i++ {
- layer.Storage(accountKey, storageKey)
- }
- }
- // With accountList and sorting
- //BenchmarkFlatten-6 50 29890856 ns/op
- //
- // Without sorting and tracking accountlist
- // BenchmarkFlatten-6 300 5511511 ns/op
- func BenchmarkFlatten(b *testing.B) {
- fill := func(parent snapshot, blocknum int) *diffLayer {
- accounts := make(map[common.Hash][]byte)
- storage := make(map[common.Hash]map[common.Hash][]byte)
- for i := 0; i < 100; i++ {
- accountKey := randomHash()
- accounts[accountKey] = randomAccount()
- accStorage := make(map[common.Hash][]byte)
- for i := 0; i < 20; i++ {
- value := make([]byte, 32)
- rand.Read(value)
- accStorage[randomHash()] = value
- }
- storage[accountKey] = accStorage
- }
- return newDiffLayer(parent, uint64(blocknum), common.Hash{}, accounts, storage)
- }
- b.ResetTimer()
- for i := 0; i < b.N; i++ {
- b.StopTimer()
- var layer snapshot
- layer = emptyLayer{}
- for i := 1; i < 128; i++ {
- layer = fill(layer, i)
- }
- b.StartTimer()
- for i := 1; i < 128; i++ {
- dl, ok := layer.(*diffLayer)
- if !ok {
- break
- }
- layer = dl.flatten()
- }
- b.StopTimer()
- }
- }
|