github.1485827954.workers.dev/ethereum/go-ethereum@v1.14.3/core/state/snapshot/disklayer.go (about) 1 // Copyright 2019 The go-ethereum Authors 2 // This file is part of the go-ethereum library. 3 // 4 // The go-ethereum library is free software: you can redistribute it and/or modify 5 // it under the terms of the GNU Lesser General Public License as published by 6 // the Free Software Foundation, either version 3 of the License, or 7 // (at your option) any later version. 8 // 9 // The go-ethereum library is distributed in the hope that it will be useful, 10 // but WITHOUT ANY WARRANTY; without even the implied warranty of 11 // MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the 12 // GNU Lesser General Public License for more details. 13 // 14 // You should have received a copy of the GNU Lesser General Public License 15 // along with the go-ethereum library. If not, see <http://www.gnu.org/licenses/>. 16 17 package snapshot 18 19 import ( 20 "bytes" 21 "sync" 22 23 "github.com/VictoriaMetrics/fastcache" 24 "github.com/ethereum/go-ethereum/common" 25 "github.com/ethereum/go-ethereum/core/rawdb" 26 "github.com/ethereum/go-ethereum/core/types" 27 "github.com/ethereum/go-ethereum/ethdb" 28 "github.com/ethereum/go-ethereum/rlp" 29 "github.com/ethereum/go-ethereum/triedb" 30 ) 31 32 // diskLayer is a low level persistent snapshot built on top of a key-value store. 33 type diskLayer struct { 34 diskdb ethdb.KeyValueStore // Key-value store containing the base snapshot 35 triedb *triedb.Database // Trie node cache for reconstruction purposes 36 cache *fastcache.Cache // Cache to avoid hitting the disk for direct access 37 38 root common.Hash // Root hash of the base snapshot 39 stale bool // Signals that the layer became stale (state progressed) 40 41 genMarker []byte // Marker for the state that's indexed during initial layer generation 42 genPending chan struct{} // Notification channel when generation is done (test synchronicity) 43 genAbort chan chan *generatorStats // Notification channel to abort generating the snapshot in this layer 44 45 lock sync.RWMutex 46 } 47 48 // Release releases underlying resources; specifically the fastcache requires 49 // Reset() in order to not leak memory. 50 // OBS: It does not invoke Close on the diskdb 51 func (dl *diskLayer) Release() error { 52 if dl.cache != nil { 53 dl.cache.Reset() 54 } 55 return nil 56 } 57 58 // Root returns root hash for which this snapshot was made. 59 func (dl *diskLayer) Root() common.Hash { 60 return dl.root 61 } 62 63 // Parent always returns nil as there's no layer below the disk. 64 func (dl *diskLayer) Parent() snapshot { 65 return nil 66 } 67 68 // Stale return whether this layer has become stale (was flattened across) or if 69 // it's still live. 70 func (dl *diskLayer) Stale() bool { 71 dl.lock.RLock() 72 defer dl.lock.RUnlock() 73 74 return dl.stale 75 } 76 77 // Account directly retrieves the account associated with a particular hash in 78 // the snapshot slim data format. 79 func (dl *diskLayer) Account(hash common.Hash) (*types.SlimAccount, error) { 80 data, err := dl.AccountRLP(hash) 81 if err != nil { 82 return nil, err 83 } 84 if len(data) == 0 { // can be both nil and []byte{} 85 return nil, nil 86 } 87 account := new(types.SlimAccount) 88 if err := rlp.DecodeBytes(data, account); err != nil { 89 panic(err) 90 } 91 return account, nil 92 } 93 94 // AccountRLP directly retrieves the account RLP associated with a particular 95 // hash in the snapshot slim data format. 96 func (dl *diskLayer) AccountRLP(hash common.Hash) ([]byte, error) { 97 dl.lock.RLock() 98 defer dl.lock.RUnlock() 99 100 // If the layer was flattened into, consider it invalid (any live reference to 101 // the original should be marked as unusable). 102 if dl.stale { 103 return nil, ErrSnapshotStale 104 } 105 // If the layer is being generated, ensure the requested hash has already been 106 // covered by the generator. 107 if dl.genMarker != nil && bytes.Compare(hash[:], dl.genMarker) > 0 { 108 return nil, ErrNotCoveredYet 109 } 110 // If we're in the disk layer, all diff layers missed 111 snapshotDirtyAccountMissMeter.Mark(1) 112 113 // Try to retrieve the account from the memory cache 114 if blob, found := dl.cache.HasGet(nil, hash[:]); found { 115 snapshotCleanAccountHitMeter.Mark(1) 116 snapshotCleanAccountReadMeter.Mark(int64(len(blob))) 117 return blob, nil 118 } 119 // Cache doesn't contain account, pull from disk and cache for later 120 blob := rawdb.ReadAccountSnapshot(dl.diskdb, hash) 121 dl.cache.Set(hash[:], blob) 122 123 snapshotCleanAccountMissMeter.Mark(1) 124 if n := len(blob); n > 0 { 125 snapshotCleanAccountWriteMeter.Mark(int64(n)) 126 } else { 127 snapshotCleanAccountInexMeter.Mark(1) 128 } 129 return blob, nil 130 } 131 132 // Storage directly retrieves the storage data associated with a particular hash, 133 // within a particular account. 134 func (dl *diskLayer) Storage(accountHash, storageHash common.Hash) ([]byte, error) { 135 dl.lock.RLock() 136 defer dl.lock.RUnlock() 137 138 // If the layer was flattened into, consider it invalid (any live reference to 139 // the original should be marked as unusable). 140 if dl.stale { 141 return nil, ErrSnapshotStale 142 } 143 key := append(accountHash[:], storageHash[:]...) 144 145 // If the layer is being generated, ensure the requested hash has already been 146 // covered by the generator. 147 if dl.genMarker != nil && bytes.Compare(key, dl.genMarker) > 0 { 148 return nil, ErrNotCoveredYet 149 } 150 // If we're in the disk layer, all diff layers missed 151 snapshotDirtyStorageMissMeter.Mark(1) 152 153 // Try to retrieve the storage slot from the memory cache 154 if blob, found := dl.cache.HasGet(nil, key); found { 155 snapshotCleanStorageHitMeter.Mark(1) 156 snapshotCleanStorageReadMeter.Mark(int64(len(blob))) 157 return blob, nil 158 } 159 // Cache doesn't contain storage slot, pull from disk and cache for later 160 blob := rawdb.ReadStorageSnapshot(dl.diskdb, accountHash, storageHash) 161 dl.cache.Set(key, blob) 162 163 snapshotCleanStorageMissMeter.Mark(1) 164 if n := len(blob); n > 0 { 165 snapshotCleanStorageWriteMeter.Mark(int64(n)) 166 } else { 167 snapshotCleanStorageInexMeter.Mark(1) 168 } 169 return blob, nil 170 } 171 172 // Update creates a new layer on top of the existing snapshot diff tree with 173 // the specified data items. Note, the maps are retained by the method to avoid 174 // copying everything. 175 func (dl *diskLayer) Update(blockHash common.Hash, destructs map[common.Hash]struct{}, accounts map[common.Hash][]byte, storage map[common.Hash]map[common.Hash][]byte) *diffLayer { 176 return newDiffLayer(dl, blockHash, destructs, accounts, storage) 177 }