github.com/muhammedhassanm/blockchain@v0.0.0-20200120143007-697261defd4d/go-ethereum-master/trie/hasher.go (about) 1 // Copyright 2016 The go-ethereum Authors 2 // This file is part of the go-ethereum library. 3 // 4 // The go-ethereum library is free software: you can redistribute it and/or modify 5 // it under the terms of the GNU Lesser General Public License as published by 6 // the Free Software Foundation, either version 3 of the License, or 7 // (at your option) any later version. 8 // 9 // The go-ethereum library is distributed in the hope that it will be useful, 10 // but WITHOUT ANY WARRANTY; without even the implied warranty of 11 // MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the 12 // GNU Lesser General Public License for more details. 13 // 14 // You should have received a copy of the GNU Lesser General Public License 15 // along with the go-ethereum library. If not, see <http://www.gnu.org/licenses/>. 16 17 package trie 18 19 import ( 20 "hash" 21 "sync" 22 23 "github.com/ethereum/go-ethereum/common" 24 "github.com/ethereum/go-ethereum/crypto/sha3" 25 "github.com/ethereum/go-ethereum/rlp" 26 ) 27 28 type hasher struct { 29 tmp sliceBuffer 30 sha keccakState 31 cachegen uint16 32 cachelimit uint16 33 onleaf LeafCallback 34 } 35 36 // keccakState wraps sha3.state. In addition to the usual hash methods, it also supports 37 // Read to get a variable amount of data from the hash state. Read is faster than Sum 38 // because it doesn't copy the internal state, but also modifies the internal state. 39 type keccakState interface { 40 hash.Hash 41 Read([]byte) (int, error) 42 } 43 44 type sliceBuffer []byte 45 46 func (b *sliceBuffer) Write(data []byte) (n int, err error) { 47 *b = append(*b, data...) 48 return len(data), nil 49 } 50 51 func (b *sliceBuffer) Reset() { 52 *b = (*b)[:0] 53 } 54 55 // hashers live in a global db. 56 var hasherPool = sync.Pool{ 57 New: func() interface{} { 58 return &hasher{ 59 tmp: make(sliceBuffer, 0, 550), // cap is as large as a full fullNode. 60 sha: sha3.NewKeccak256().(keccakState), 61 } 62 }, 63 } 64 65 func newHasher(cachegen, cachelimit uint16, onleaf LeafCallback) *hasher { 66 h := hasherPool.Get().(*hasher) 67 h.cachegen, h.cachelimit, h.onleaf = cachegen, cachelimit, onleaf 68 return h 69 } 70 71 func returnHasherToPool(h *hasher) { 72 hasherPool.Put(h) 73 } 74 75 // hash collapses a node down into a hash node, also returning a copy of the 76 // original node initialized with the computed hash to replace the original one. 77 func (h *hasher) hash(n node, db *Database, force bool) (node, node, error) { 78 // If we're not storing the node, just hashing, use available cached data 79 if hash, dirty := n.cache(); hash != nil { 80 if db == nil { 81 return hash, n, nil 82 } 83 if n.canUnload(h.cachegen, h.cachelimit) { 84 // Unload the node from cache. All of its subnodes will have a lower or equal 85 // cache generation number. 86 cacheUnloadCounter.Inc(1) 87 return hash, hash, nil 88 } 89 if !dirty { 90 return hash, n, nil 91 } 92 } 93 // Trie not processed yet or needs storage, walk the children 94 collapsed, cached, err := h.hashChildren(n, db) 95 if err != nil { 96 return hashNode{}, n, err 97 } 98 hashed, err := h.store(collapsed, db, force) 99 if err != nil { 100 return hashNode{}, n, err 101 } 102 // Cache the hash of the node for later reuse and remove 103 // the dirty flag in commit mode. It's fine to assign these values directly 104 // without copying the node first because hashChildren copies it. 105 cachedHash, _ := hashed.(hashNode) 106 switch cn := cached.(type) { 107 case *shortNode: 108 cn.flags.hash = cachedHash 109 if db != nil { 110 cn.flags.dirty = false 111 } 112 case *fullNode: 113 cn.flags.hash = cachedHash 114 if db != nil { 115 cn.flags.dirty = false 116 } 117 } 118 return hashed, cached, nil 119 } 120 121 // hashChildren replaces the children of a node with their hashes if the encoded 122 // size of the child is larger than a hash, returning the collapsed node as well 123 // as a replacement for the original node with the child hashes cached in. 124 func (h *hasher) hashChildren(original node, db *Database) (node, node, error) { 125 var err error 126 127 switch n := original.(type) { 128 case *shortNode: 129 // Hash the short node's child, caching the newly hashed subtree 130 collapsed, cached := n.copy(), n.copy() 131 collapsed.Key = hexToCompact(n.Key) 132 cached.Key = common.CopyBytes(n.Key) 133 134 if _, ok := n.Val.(valueNode); !ok { 135 collapsed.Val, cached.Val, err = h.hash(n.Val, db, false) 136 if err != nil { 137 return original, original, err 138 } 139 } 140 return collapsed, cached, nil 141 142 case *fullNode: 143 // Hash the full node's children, caching the newly hashed subtrees 144 collapsed, cached := n.copy(), n.copy() 145 146 for i := 0; i < 16; i++ { 147 if n.Children[i] != nil { 148 collapsed.Children[i], cached.Children[i], err = h.hash(n.Children[i], db, false) 149 if err != nil { 150 return original, original, err 151 } 152 } 153 } 154 cached.Children[16] = n.Children[16] 155 return collapsed, cached, nil 156 157 default: 158 // Value and hash nodes don't have children so they're left as were 159 return n, original, nil 160 } 161 } 162 163 // store hashes the node n and if we have a storage layer specified, it writes 164 // the key/value pair to it and tracks any node->child references as well as any 165 // node->external trie references. 166 func (h *hasher) store(n node, db *Database, force bool) (node, error) { 167 // Don't store hashes or empty nodes. 168 if _, isHash := n.(hashNode); n == nil || isHash { 169 return n, nil 170 } 171 // Generate the RLP encoding of the node 172 h.tmp.Reset() 173 if err := rlp.Encode(&h.tmp, n); err != nil { 174 panic("encode error: " + err.Error()) 175 } 176 if len(h.tmp) < 32 && !force { 177 return n, nil // Nodes smaller than 32 bytes are stored inside their parent 178 } 179 // Larger nodes are replaced by their hash and stored in the database. 180 hash, _ := n.cache() 181 if hash == nil { 182 hash = h.makeHashNode(h.tmp) 183 } 184 185 if db != nil { 186 // We are pooling the trie nodes into an intermediate memory cache 187 hash := common.BytesToHash(hash) 188 189 db.lock.Lock() 190 db.insert(hash, h.tmp, n) 191 db.lock.Unlock() 192 193 // Track external references from account->storage trie 194 if h.onleaf != nil { 195 switch n := n.(type) { 196 case *shortNode: 197 if child, ok := n.Val.(valueNode); ok { 198 h.onleaf(child, hash) 199 } 200 case *fullNode: 201 for i := 0; i < 16; i++ { 202 if child, ok := n.Children[i].(valueNode); ok { 203 h.onleaf(child, hash) 204 } 205 } 206 } 207 } 208 } 209 return hash, nil 210 } 211 212 func (h *hasher) makeHashNode(data []byte) hashNode { 213 n := make(hashNode, h.sha.Size()) 214 h.sha.Reset() 215 h.sha.Write(data) 216 h.sha.Read(n) 217 return n 218 }