// Copyright 2019 The go-ethereum Authors // This file is part of the go-ethereum library. // // The go-ethereum library is free software: you can redistribute it and/or modify // it under the terms of the GNU Lesser General Public License as published by // the Free Software Foundation, either version 3 of the License, or // (at your option) any later version. // // The go-ethereum library is distributed in the hope that it will be useful, // but WITHOUT ANY WARRANTY; without even the implied warranty of // MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the // GNU Lesser General Public License for more details. // // You should have received a copy of the GNU Lesser General Public License // along with the go-ethereum library. If not, see . package snapshot import ( "sync" "github.com/allegro/bigcache" "github.com/ethereum/go-ethereum/common" "github.com/ethereum/go-ethereum/core/rawdb" "github.com/ethereum/go-ethereum/ethdb" "github.com/ethereum/go-ethereum/rlp" ) // diskLayer is a low level persistent snapshot built on top of a key-value store. type diskLayer struct { journal string // Path of the snapshot journal to use on shutdown db ethdb.KeyValueStore // Key-value store containing the base snapshot cache *bigcache.BigCache // Cache to avoid hitting the disk for direct access root common.Hash // Root hash of the base snapshot stale bool // Signals that the layer became stale (state progressed) lock sync.RWMutex } // Root returns root hash for which this snapshot was made. func (dl *diskLayer) Root() common.Hash { return dl.root } // Stale return whether this layer has become stale (was flattened across) or if // it's still live. func (dl *diskLayer) Stale() bool { dl.lock.RLock() defer dl.lock.RUnlock() return dl.stale } // Account directly retrieves the account associated with a particular hash in // the snapshot slim data format. func (dl *diskLayer) Account(hash common.Hash) (*Account, error) { data, err := dl.AccountRLP(hash) if err != nil { return nil, err } if len(data) == 0 { // can be both nil and []byte{} return nil, nil } account := new(Account) if err := rlp.DecodeBytes(data, account); err != nil { panic(err) } return account, nil } // AccountRLP directly retrieves the account RLP associated with a particular // hash in the snapshot slim data format. func (dl *diskLayer) AccountRLP(hash common.Hash) ([]byte, error) { dl.lock.RLock() defer dl.lock.RUnlock() // If the layer was flattened into, consider it invalid (any live reference to // the original should be marked as unusable). if dl.stale { return nil, ErrSnapshotStale } key := string(hash[:]) // Try to retrieve the account from the memory cache if blob, err := dl.cache.Get(key); err == nil { snapshotCleanHitMeter.Mark(1) snapshotCleanReadMeter.Mark(int64(len(blob))) return blob, nil } // Cache doesn't contain account, pull from disk and cache for later blob := rawdb.ReadAccountSnapshot(dl.db, hash) dl.cache.Set(key, blob) snapshotCleanMissMeter.Mark(1) snapshotCleanWriteMeter.Mark(int64(len(blob))) return blob, nil } // Storage directly retrieves the storage data associated with a particular hash, // within a particular account. func (dl *diskLayer) Storage(accountHash, storageHash common.Hash) ([]byte, error) { dl.lock.RLock() defer dl.lock.RUnlock() // If the layer was flattened into, consider it invalid (any live reference to // the original should be marked as unusable). if dl.stale { return nil, ErrSnapshotStale } key := string(append(accountHash[:], storageHash[:]...)) // Try to retrieve the storage slot from the memory cache if blob, err := dl.cache.Get(key); err == nil { snapshotCleanHitMeter.Mark(1) snapshotCleanReadMeter.Mark(int64(len(blob))) return blob, nil } // Cache doesn't contain storage slot, pull from disk and cache for later blob := rawdb.ReadStorageSnapshot(dl.db, accountHash, storageHash) dl.cache.Set(key, blob) snapshotCleanMissMeter.Mark(1) snapshotCleanWriteMeter.Mark(int64(len(blob))) return blob, nil } // Update creates a new layer on top of the existing snapshot diff tree with // the specified data items. Note, the maps are retained by the method to avoid // copying everything. func (dl *diskLayer) Update(blockHash common.Hash, accounts map[common.Hash][]byte, storage map[common.Hash]map[common.Hash][]byte) *diffLayer { return newDiffLayer(dl, blockHash, accounts, storage) } // Journal commits an entire diff hierarchy to disk into a single journal file. func (dl *diskLayer) Journal() error { // There's no journalling a disk layer return nil }