// Copyright 2020 The go-ethereum Authors
// This file is part of go-ethereum.
//
// go-ethereum is free software: you can redistribute it and/or modify
// it under the terms of the GNU General Public License as published by
// the Free Software Foundation, either version 3 of the License, or
// (at your option) any later version.
//
// go-ethereum is distributed in the hope that it will be useful,
// but WITHOUT ANY WARRANTY; without even the implied warranty of
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
// GNU General Public License for more details.
//
// You should have received a copy of the GNU General Public License
// along with go-ethereum. If not, see <http://www.gnu.org/licenses/>.
package main
import (
"bytes"
"errors"
"fmt"
"os"
"os/signal"
"path/filepath"
"sort"
"strconv"
"strings"
"syscall"
"time"
"github.com/ethereum/go-ethereum/cmd/utils"
"github.com/ethereum/go-ethereum/common"
"github.com/ethereum/go-ethereum/common/hexutil"
"github.com/ethereum/go-ethereum/console/prompt"
"github.com/ethereum/go-ethereum/core/rawdb"
"github.com/ethereum/go-ethereum/core/state/snapshot"
"github.com/ethereum/go-ethereum/core/types"
"github.com/ethereum/go-ethereum/ethdb"
"github.com/ethereum/go-ethereum/log"
"github.com/ethereum/go-ethereum/trie"
"github.com/olekukonko/tablewriter"
"gopkg.in/urfave/cli.v1"
)
var (
removedbCommand = cli . Command {
Action : utils . MigrateFlags ( removeDB ) ,
Name : "removedb" ,
Usage : "Remove blockchain and state databases" ,
ArgsUsage : "" ,
Flags : [ ] cli . Flag {
utils . DataDirFlag ,
} ,
Category : "DATABASE COMMANDS" ,
Description : `
Remove blockchain and state databases ` ,
}
dbCommand = cli . Command {
Name : "db" ,
Usage : "Low level database operations" ,
ArgsUsage : "" ,
Category : "DATABASE COMMANDS" ,
Subcommands : [ ] cli . Command {
dbInspectCmd ,
dbStatCmd ,
dbCompactCmd ,
dbGetCmd ,
dbDeleteCmd ,
dbPutCmd ,
dbGetSlotsCmd ,
dbDumpFreezerIndex ,
dbImportCmd ,
dbExportCmd ,
dbMetadataCmd ,
dbMigrateFreezerCmd ,
} ,
}
dbInspectCmd = cli . Command {
Action : utils . MigrateFlags ( inspect ) ,
Name : "inspect" ,
ArgsUsage : "<prefix> <start>" ,
Flags : [ ] cli . Flag {
utils . DataDirFlag ,
utils . AncientFlag ,
utils . SyncModeFlag ,
utils . MainnetFlag ,
utils . RopstenFlag ,
utils . SepoliaFlag ,
utils . RinkebyFlag ,
utils . GoerliFlag ,
} ,
Usage : "Inspect the storage size for each type of data in the database" ,
Description : ` This commands iterates the entire database. If the optional 'prefix' and 'start' arguments are provided, then the iteration is limited to the given subset of data. ` ,
}
dbStatCmd = cli . Command {
Action : utils . MigrateFlags ( dbStats ) ,
Name : "stats" ,
Usage : "Print leveldb statistics" ,
Flags : [ ] cli . Flag {
utils . DataDirFlag ,
utils . SyncModeFlag ,
utils . MainnetFlag ,
utils . RopstenFlag ,
utils . SepoliaFlag ,
utils . RinkebyFlag ,
utils . GoerliFlag ,
} ,
}
dbCompactCmd = cli . Command {
Action : utils . MigrateFlags ( dbCompact ) ,
Name : "compact" ,
Usage : "Compact leveldb database. WARNING: May take a very long time" ,
Flags : [ ] cli . Flag {
utils . DataDirFlag ,
utils . SyncModeFlag ,
utils . MainnetFlag ,
utils . RopstenFlag ,
utils . SepoliaFlag ,
utils . RinkebyFlag ,
utils . GoerliFlag ,
utils . CacheFlag ,
utils . CacheDatabaseFlag ,
} ,
Description : ` This command performs a database compaction .
WARNING : This operation may take a very long time to finish , and may cause database
corruption if it is aborted during execution ' ! ` ,
}
dbGetCmd = cli . Command {
Action : utils . MigrateFlags ( dbGet ) ,
Name : "get" ,
Usage : "Show the value of a database key" ,
ArgsUsage : "<hex-encoded key>" ,
Flags : [ ] cli . Flag {
utils . DataDirFlag ,
utils . SyncModeFlag ,
utils . MainnetFlag ,
utils . RopstenFlag ,
utils . SepoliaFlag ,
utils . RinkebyFlag ,
utils . GoerliFlag ,
} ,
Description : "This command looks up the specified database key from the database." ,
}
dbDeleteCmd = cli . Command {
Action : utils . MigrateFlags ( dbDelete ) ,
Name : "delete" ,
Usage : "Delete a database key (WARNING: may corrupt your database)" ,
ArgsUsage : "<hex-encoded key>" ,
Flags : [ ] cli . Flag {
utils . DataDirFlag ,
utils . SyncModeFlag ,
utils . MainnetFlag ,
utils . RopstenFlag ,
utils . SepoliaFlag ,
utils . RinkebyFlag ,
utils . GoerliFlag ,
} ,
Description : ` This command deletes the specified database key from the database .
WARNING : This is a low - level operation which may cause database corruption ! ` ,
}
dbPutCmd = cli . Command {
Action : utils . MigrateFlags ( dbPut ) ,
Name : "put" ,
Usage : "Set the value of a database key (WARNING: may corrupt your database)" ,
ArgsUsage : "<hex-encoded key> <hex-encoded value>" ,
Flags : [ ] cli . Flag {
utils . DataDirFlag ,
utils . SyncModeFlag ,
utils . MainnetFlag ,
utils . RopstenFlag ,
utils . SepoliaFlag ,
utils . RinkebyFlag ,
utils . GoerliFlag ,
} ,
Description : ` This command sets a given database key to the given value .
WARNING : This is a low - level operation which may cause database corruption ! ` ,
}
dbGetSlotsCmd = cli . Command {
Action : utils . MigrateFlags ( dbDumpTrie ) ,
Name : "dumptrie" ,
Usage : "Show the storage key/values of a given storage trie" ,
ArgsUsage : "<hex-encoded storage trie root> <hex-encoded start (optional)> <int max elements (optional)>" ,
Flags : [ ] cli . Flag {
utils . DataDirFlag ,
utils . SyncModeFlag ,
utils . MainnetFlag ,
utils . RopstenFlag ,
utils . SepoliaFlag ,
utils . RinkebyFlag ,
utils . GoerliFlag ,
} ,
Description : "This command looks up the specified database key from the database." ,
}
dbDumpFreezerIndex = cli . Command {
Action : utils . MigrateFlags ( freezerInspect ) ,
Name : "freezer-index" ,
Usage : "Dump out the index of a given freezer type" ,
ArgsUsage : "<type> <start (int)> <end (int)>" ,
Flags : [ ] cli . Flag {
utils . DataDirFlag ,
utils . SyncModeFlag ,
utils . MainnetFlag ,
utils . RopstenFlag ,
utils . SepoliaFlag ,
utils . RinkebyFlag ,
utils . GoerliFlag ,
} ,
Description : "This command displays information about the freezer index." ,
}
dbImportCmd = cli . Command {
Action : utils . MigrateFlags ( importLDBdata ) ,
Name : "import" ,
Usage : "Imports leveldb-data from an exported RLP dump." ,
ArgsUsage : "<dumpfile> <start (optional)" ,
Flags : [ ] cli . Flag {
utils . DataDirFlag ,
utils . SyncModeFlag ,
utils . MainnetFlag ,
utils . RopstenFlag ,
utils . RinkebyFlag ,
utils . GoerliFlag ,
} ,
Description : "The import command imports the specific chain data from an RLP encoded stream." ,
}
dbExportCmd = cli . Command {
Action : utils . MigrateFlags ( exportChaindata ) ,
Name : "export" ,
Usage : "Exports the chain data into an RLP dump. If the <dumpfile> has .gz suffix, gzip compression will be used." ,
ArgsUsage : "<type> <dumpfile>" ,
Flags : [ ] cli . Flag {
utils . DataDirFlag ,
utils . SyncModeFlag ,
utils . MainnetFlag ,
utils . RopstenFlag ,
utils . RinkebyFlag ,
utils . GoerliFlag ,
} ,
Description : "Exports the specified chain data to an RLP encoded stream, optionally gzip-compressed." ,
}
dbMetadataCmd = cli . Command {
Action : utils . MigrateFlags ( showMetaData ) ,
Name : "metadata" ,
Usage : "Shows metadata about the chain status." ,
Flags : [ ] cli . Flag {
utils . DataDirFlag ,
utils . SyncModeFlag ,
utils . MainnetFlag ,
utils . RopstenFlag ,
utils . SepoliaFlag ,
utils . RinkebyFlag ,
utils . GoerliFlag ,
} ,
Description : "Shows metadata about the chain status." ,
}
dbMigrateFreezerCmd = cli . Command {
Action : utils . MigrateFlags ( freezerMigrate ) ,
Name : "freezer-migrate" ,
Usage : "Migrate legacy parts of the freezer. (WARNING: may take a long time)" ,
ArgsUsage : "" ,
Flags : [ ] cli . Flag {
utils . DataDirFlag ,
utils . SyncModeFlag ,
utils . MainnetFlag ,
utils . RopstenFlag ,
utils . SepoliaFlag ,
utils . RinkebyFlag ,
utils . GoerliFlag ,
} ,
Description : ` The freezer - migrate command checks your database for receipts in a legacy format and updates those .
WARNING : please back - up the receipt files in your ancients before running this command . ` ,
}
)
func removeDB ( ctx * cli . Context ) error {
stack , config := makeConfigNode ( ctx )
// Remove the full node state database
path := stack . ResolvePath ( "chaindata" )
if common . FileExist ( path ) {
confirmAndRemoveDB ( path , "full node state database" )
} else {
log . Info ( "Full node state database missing" , "path" , path )
}
// Remove the full node ancient database
path = config . Eth . DatabaseFreezer
switch {
case path == "" :
path = filepath . Join ( stack . ResolvePath ( "chaindata" ) , "ancient" )
case ! filepath . IsAbs ( path ) :
path = config . Node . ResolvePath ( path )
}
if common . FileExist ( path ) {
confirmAndRemoveDB ( path , "full node ancient database" )
} else {
log . Info ( "Full node ancient database missing" , "path" , path )
}
// Remove the light node database
path = stack . ResolvePath ( "lightchaindata" )
if common . FileExist ( path ) {
confirmAndRemoveDB ( path , "light node database" )
} else {
log . Info ( "Light node database missing" , "path" , path )
}
return nil
}
// confirmAndRemoveDB prompts the user for a last confirmation and removes the
// folder if accepted.
func confirmAndRemoveDB ( database string , kind string ) {
confirm , err := prompt . Stdin . PromptConfirm ( fmt . Sprintf ( "Remove %s (%s)?" , kind , database ) )
switch {
case err != nil :
utils . Fatalf ( "%v" , err )
case ! confirm :
log . Info ( "Database deletion skipped" , "path" , database )
default :
start := time . Now ( )
filepath . Walk ( database , func ( path string , info os . FileInfo , err error ) error {
// If we're at the top level folder, recurse into
if path == database {
return nil
}
// Delete all the files, but not subfolders
if ! info . IsDir ( ) {
os . Remove ( path )
return nil
}
return filepath . SkipDir
} )
log . Info ( "Database successfully deleted" , "path" , database , "elapsed" , common . PrettyDuration ( time . Since ( start ) ) )
}
}
func inspect ( ctx * cli . Context ) error {
var (
prefix [ ] byte
start [ ] byte
)
if ctx . NArg ( ) > 2 {
return fmt . Errorf ( "Max 2 arguments: %v" , ctx . Command . ArgsUsage )
}
if ctx . NArg ( ) >= 1 {
if d , err := hexutil . Decode ( ctx . Args ( ) . Get ( 0 ) ) ; err != nil {
return fmt . Errorf ( "failed to hex-decode 'prefix': %v" , err )
} else {
prefix = d
}
}
if ctx . NArg ( ) >= 2 {
if d , err := hexutil . Decode ( ctx . Args ( ) . Get ( 1 ) ) ; err != nil {
return fmt . Errorf ( "failed to hex-decode 'start': %v" , err )
} else {
start = d
}
}
stack , _ := makeConfigNode ( ctx )
defer stack . Close ( )
db := utils . MakeChainDatabase ( ctx , stack , true )
defer db . Close ( )
return rawdb . InspectDatabase ( db , prefix , start )
}
func showLeveldbStats ( db ethdb . Stater ) {
if stats , err := db . Stat ( "leveldb.stats" ) ; err != nil {
log . Warn ( "Failed to read database stats" , "error" , err )
} else {
fmt . Println ( stats )
}
if ioStats , err := db . Stat ( "leveldb.iostats" ) ; err != nil {
log . Warn ( "Failed to read database iostats" , "error" , err )
} else {
fmt . Println ( ioStats )
}
}
func dbStats ( ctx * cli . Context ) error {
stack , _ := makeConfigNode ( ctx )
defer stack . Close ( )
db := utils . MakeChainDatabase ( ctx , stack , true )
defer db . Close ( )
showLeveldbStats ( db )
return nil
}
func dbCompact ( ctx * cli . Context ) error {
stack , _ := makeConfigNode ( ctx )
defer stack . Close ( )
db := utils . MakeChainDatabase ( ctx , stack , false )
defer db . Close ( )
log . Info ( "Stats before compaction" )
showLeveldbStats ( db )
log . Info ( "Triggering compaction" )
if err := db . Compact ( nil , nil ) ; err != nil {
log . Info ( "Compact err" , "error" , err )
return err
}
log . Info ( "Stats after compaction" )
showLeveldbStats ( db )
return nil
}
// dbGet shows the value of a given database key
func dbGet ( ctx * cli . Context ) error {
if ctx . NArg ( ) != 1 {
return fmt . Errorf ( "required arguments: %v" , ctx . Command . ArgsUsage )
}
stack , _ := makeConfigNode ( ctx )
defer stack . Close ( )
db := utils . MakeChainDatabase ( ctx , stack , true )
defer db . Close ( )
key , err := parseHexOrString ( ctx . Args ( ) . Get ( 0 ) )
if err != nil {
log . Info ( "Could not decode the key" , "error" , err )
return err
}
data , err := db . Get ( key )
if err != nil {
log . Info ( "Get operation failed" , "key" , fmt . Sprintf ( "0x%#x" , key ) , "error" , err )
return err
}
fmt . Printf ( "key %#x: %#x\n" , key , data )
return nil
}
// dbDelete deletes a key from the database
func dbDelete ( ctx * cli . Context ) error {
if ctx . NArg ( ) != 1 {
return fmt . Errorf ( "required arguments: %v" , ctx . Command . ArgsUsage )
}
stack , _ := makeConfigNode ( ctx )
defer stack . Close ( )
db := utils . MakeChainDatabase ( ctx , stack , false )
defer db . Close ( )
key , err := parseHexOrString ( ctx . Args ( ) . Get ( 0 ) )
if err != nil {
log . Info ( "Could not decode the key" , "error" , err )
return err
}
data , err := db . Get ( key )
if err == nil {
fmt . Printf ( "Previous value: %#x\n" , data )
}
if err = db . Delete ( key ) ; err != nil {
log . Info ( "Delete operation returned an error" , "key" , fmt . Sprintf ( "0x%#x" , key ) , "error" , err )
return err
}
return nil
}
// dbPut overwrite a value in the database
func dbPut ( ctx * cli . Context ) error {
if ctx . NArg ( ) != 2 {
return fmt . Errorf ( "required arguments: %v" , ctx . Command . ArgsUsage )
}
stack , _ := makeConfigNode ( ctx )
defer stack . Close ( )
db := utils . MakeChainDatabase ( ctx , stack , false )
defer db . Close ( )
var (
key [ ] byte
value [ ] byte
data [ ] byte
err error
)
key , err = parseHexOrString ( ctx . Args ( ) . Get ( 0 ) )
if err != nil {
log . Info ( "Could not decode the key" , "error" , err )
return err
}
value , err = hexutil . Decode ( ctx . Args ( ) . Get ( 1 ) )
if err != nil {
log . Info ( "Could not decode the value" , "error" , err )
return err
}
data , err = db . Get ( key )
if err == nil {
fmt . Printf ( "Previous value: %#x\n" , data )
}
return db . Put ( key , value )
}
// dbDumpTrie shows the key-value slots of a given storage trie
func dbDumpTrie ( ctx * cli . Context ) error {
if ctx . NArg ( ) < 1 {
return fmt . Errorf ( "required arguments: %v" , ctx . Command . ArgsUsage )
}
stack , _ := makeConfigNode ( ctx )
defer stack . Close ( )
db := utils . MakeChainDatabase ( ctx , stack , true )
defer db . Close ( )
var (
root [ ] byte
start [ ] byte
max = int64 ( - 1 )
err error
)
if root , err = hexutil . Decode ( ctx . Args ( ) . Get ( 0 ) ) ; err != nil {
log . Info ( "Could not decode the root" , "error" , err )
return err
}
stRoot := common . BytesToHash ( root )
if ctx . NArg ( ) >= 2 {
if start , err = hexutil . Decode ( ctx . Args ( ) . Get ( 1 ) ) ; err != nil {
log . Info ( "Could not decode the seek position" , "error" , err )
return err
}
}
if ctx . NArg ( ) >= 3 {
if max , err = strconv . ParseInt ( ctx . Args ( ) . Get ( 2 ) , 10 , 64 ) ; err != nil {
log . Info ( "Could not decode the max count" , "error" , err )
return err
}
}
theTrie , err := trie . New ( stRoot , trie . NewDatabase ( db ) )
if err != nil {
return err
}
var count int64
it := trie . NewIterator ( theTrie . NodeIterator ( start ) )
for it . Next ( ) {
if max > 0 && count == max {
fmt . Printf ( "Exiting after %d values\n" , count )
break
}
fmt . Printf ( " %d. key %#x: %#x\n" , count , it . Key , it . Value )
count ++
}
return it . Err
}
func freezerInspect ( ctx * cli . Context ) error {
var (
start , end int64
disableSnappy bool
err error
)
if ctx . NArg ( ) < 3 {
return fmt . Errorf ( "required arguments: %v" , ctx . Command . ArgsUsage )
}
kind := ctx . Args ( ) . Get ( 0 )
if noSnap , ok := rawdb . FreezerNoSnappy [ kind ] ; ! ok {
var options [ ] string
for opt := range rawdb . FreezerNoSnappy {
options = append ( options , opt )
}
sort . Strings ( options )
return fmt . Errorf ( "Could read freezer-type '%v'. Available options: %v" , kind , options )
} else {
disableSnappy = noSnap
}
if start , err = strconv . ParseInt ( ctx . Args ( ) . Get ( 1 ) , 10 , 64 ) ; err != nil {
log . Info ( "Could read start-param" , "error" , err )
return err
}
if end , err = strconv . ParseInt ( ctx . Args ( ) . Get ( 2 ) , 10 , 64 ) ; err != nil {
log . Info ( "Could read count param" , "error" , err )
return err
}
stack , _ := makeConfigNode ( ctx )
defer stack . Close ( )
path := filepath . Join ( stack . ResolvePath ( "chaindata" ) , "ancient" )
log . Info ( "Opening freezer" , "location" , path , "name" , kind )
if f , err := rawdb . NewFreezerTable ( path , kind , disableSnappy , true ) ; err != nil {
return err
} else {
f . DumpIndex ( start , end )
}
return nil
}
// ParseHexOrString tries to hexdecode b, but if the prefix is missing, it instead just returns the raw bytes
func parseHexOrString ( str string ) ( [ ] byte , error ) {
b , err := hexutil . Decode ( str )
if errors . Is ( err , hexutil . ErrMissingPrefix ) {
return [ ] byte ( str ) , nil
}
return b , err
}
func importLDBdata ( ctx * cli . Context ) error {
start := 0
switch ctx . NArg ( ) {
case 1 :
break
case 2 :
s , err := strconv . Atoi ( ctx . Args ( ) . Get ( 1 ) )
if err != nil {
return fmt . Errorf ( "second arg must be an integer: %v" , err )
}
start = s
default :
return fmt . Errorf ( "required arguments: %v" , ctx . Command . ArgsUsage )
}
var (
fName = ctx . Args ( ) . Get ( 0 )
stack , _ = makeConfigNode ( ctx )
interrupt = make ( chan os . Signal , 1 )
stop = make ( chan struct { } )
)
defer stack . Close ( )
signal . Notify ( interrupt , syscall . SIGINT , syscall . SIGTERM )
defer signal . Stop ( interrupt )
defer close ( interrupt )
go func ( ) {
if _ , ok := <- interrupt ; ok {
log . Info ( "Interrupted during ldb import, stopping at next batch" )
}
close ( stop )
} ( )
db := utils . MakeChainDatabase ( ctx , stack , false )
return utils . ImportLDBData ( db , fName , int64 ( start ) , stop )
}
type preimageIterator struct {
iter ethdb . Iterator
}
func ( iter * preimageIterator ) Next ( ) ( byte , [ ] byte , [ ] byte , bool ) {
for iter . iter . Next ( ) {
key := iter . iter . Key ( )
if bytes . HasPrefix ( key , rawdb . PreimagePrefix ) && len ( key ) == ( len ( rawdb . PreimagePrefix ) + common . HashLength ) {
return utils . OpBatchAdd , key , iter . iter . Value ( ) , true
}
}
return 0 , nil , nil , false
}
func ( iter * preimageIterator ) Release ( ) {
iter . iter . Release ( )
}
type snapshotIterator struct {
init bool
account ethdb . Iterator
storage ethdb . Iterator
}
func ( iter * snapshotIterator ) Next ( ) ( byte , [ ] byte , [ ] byte , bool ) {
if ! iter . init {
iter . init = true
return utils . OpBatchDel , rawdb . SnapshotRootKey , nil , true
}
for iter . account . Next ( ) {
key := iter . account . Key ( )
if bytes . HasPrefix ( key , rawdb . SnapshotAccountPrefix ) && len ( key ) == ( len ( rawdb . SnapshotAccountPrefix ) + common . HashLength ) {
return utils . OpBatchAdd , key , iter . account . Value ( ) , true
}
}
for iter . storage . Next ( ) {
key := iter . storage . Key ( )
if bytes . HasPrefix ( key , rawdb . SnapshotStoragePrefix ) && len ( key ) == ( len ( rawdb . SnapshotStoragePrefix ) + 2 * common . HashLength ) {
return utils . OpBatchAdd , key , iter . storage . Value ( ) , true
}
}
return 0 , nil , nil , false
}
func ( iter * snapshotIterator ) Release ( ) {
iter . account . Release ( )
iter . storage . Release ( )
}
// chainExporters defines the export scheme for all exportable chain data.
var chainExporters = map [ string ] func ( db ethdb . Database ) utils . ChainDataIterator {
"preimage" : func ( db ethdb . Database ) utils . ChainDataIterator {
iter := db . NewIterator ( rawdb . PreimagePrefix , nil )
return & preimageIterator { iter : iter }
} ,
"snapshot" : func ( db ethdb . Database ) utils . ChainDataIterator {
account := db . NewIterator ( rawdb . SnapshotAccountPrefix , nil )
storage := db . NewIterator ( rawdb . SnapshotStoragePrefix , nil )
return & snapshotIterator { account : account , storage : storage }
} ,
}
func exportChaindata ( ctx * cli . Context ) error {
if ctx . NArg ( ) < 2 {
return fmt . Errorf ( "required arguments: %v" , ctx . Command . ArgsUsage )
}
// Parse the required chain data type, make sure it's supported.
kind := ctx . Args ( ) . Get ( 0 )
kind = strings . ToLower ( strings . Trim ( kind , " " ) )
exporter , ok := chainExporters [ kind ]
if ! ok {
var kinds [ ] string
for kind := range chainExporters {
kinds = append ( kinds , kind )
}
return fmt . Errorf ( "invalid data type %s, supported types: %s" , kind , strings . Join ( kinds , ", " ) )
}
var (
stack , _ = makeConfigNode ( ctx )
interrupt = make ( chan os . Signal , 1 )
stop = make ( chan struct { } )
)
defer stack . Close ( )
signal . Notify ( interrupt , syscall . SIGINT , syscall . SIGTERM )
defer signal . Stop ( interrupt )
defer close ( interrupt )
go func ( ) {
if _ , ok := <- interrupt ; ok {
log . Info ( "Interrupted during db export, stopping at next batch" )
}
close ( stop )
} ( )
db := utils . MakeChainDatabase ( ctx , stack , true )
return utils . ExportChaindata ( ctx . Args ( ) . Get ( 1 ) , kind , exporter ( db ) , stop )
}
func showMetaData ( ctx * cli . Context ) error {
stack , _ := makeConfigNode ( ctx )
defer stack . Close ( )
db := utils . MakeChainDatabase ( ctx , stack , true )
ancients , err := db . Ancients ( )
if err != nil {
fmt . Fprintf ( os . Stderr , "Error accessing ancients: %v" , err )
}
pp := func ( val * uint64 ) string {
if val == nil {
return "<nil>"
}
return fmt . Sprintf ( "%d (0x%x)" , * val , * val )
}
data := [ ] [ ] string {
{ "databaseVersion" , pp ( rawdb . ReadDatabaseVersion ( db ) ) } ,
{ "headBlockHash" , fmt . Sprintf ( "%v" , rawdb . ReadHeadBlockHash ( db ) ) } ,
{ "headFastBlockHash" , fmt . Sprintf ( "%v" , rawdb . ReadHeadFastBlockHash ( db ) ) } ,
{ "headHeaderHash" , fmt . Sprintf ( "%v" , rawdb . ReadHeadHeaderHash ( db ) ) } }
if b := rawdb . ReadHeadBlock ( db ) ; b != nil {
data = append ( data , [ ] string { "headBlock.Hash" , fmt . Sprintf ( "%v" , b . Hash ( ) ) } )
data = append ( data , [ ] string { "headBlock.Root" , fmt . Sprintf ( "%v" , b . Root ( ) ) } )
data = append ( data , [ ] string { "headBlock.Number" , fmt . Sprintf ( "%d (0x%x)" , b . Number ( ) , b . Number ( ) ) } )
}
if b := rawdb . ReadSkeletonSyncStatus ( db ) ; b != nil {
data = append ( data , [ ] string { "SkeletonSyncStatus" , string ( b ) } )
}
if h := rawdb . ReadHeadHeader ( db ) ; h != nil {
data = append ( data , [ ] string { "headHeader.Hash" , fmt . Sprintf ( "%v" , h . Hash ( ) ) } )
data = append ( data , [ ] string { "headHeader.Root" , fmt . Sprintf ( "%v" , h . Root ) } )
data = append ( data , [ ] string { "headHeader.Number" , fmt . Sprintf ( "%d (0x%x)" , h . Number , h . Number ) } )
}
data = append ( data , [ ] [ ] string { { "frozen" , fmt . Sprintf ( "%d items" , ancients ) } ,
{ "lastPivotNumber" , pp ( rawdb . ReadLastPivotNumber ( db ) ) } ,
{ "len(snapshotSyncStatus)" , fmt . Sprintf ( "%d bytes" , len ( rawdb . ReadSnapshotSyncStatus ( db ) ) ) } ,
{ "snapshotGenerator" , snapshot . ParseGeneratorStatus ( rawdb . ReadSnapshotGenerator ( db ) ) } ,
{ "snapshotDisabled" , fmt . Sprintf ( "%v" , rawdb . ReadSnapshotDisabled ( db ) ) } ,
{ "snapshotJournal" , fmt . Sprintf ( "%d bytes" , len ( rawdb . ReadSnapshotJournal ( db ) ) ) } ,
{ "snapshotRecoveryNumber" , pp ( rawdb . ReadSnapshotRecoveryNumber ( db ) ) } ,
{ "snapshotRoot" , fmt . Sprintf ( "%v" , rawdb . ReadSnapshotRoot ( db ) ) } ,
{ "txIndexTail" , pp ( rawdb . ReadTxIndexTail ( db ) ) } ,
{ "fastTxLookupLimit" , pp ( rawdb . ReadFastTxLookupLimit ( db ) ) } ,
} ... )
table := tablewriter . NewWriter ( os . Stdout )
table . SetHeader ( [ ] string { "Field" , "Value" } )
table . AppendBulk ( data )
table . Render ( )
return nil
}
func freezerMigrate ( ctx * cli . Context ) error {
stack , _ := makeConfigNode ( ctx )
defer stack . Close ( )
db := utils . MakeChainDatabase ( ctx , stack , false )
defer db . Close ( )
// Check first block for legacy receipt format
numAncients , err := db . Ancients ( )
if err != nil {
return err
}
if numAncients < 1 {
log . Info ( "No receipts in freezer to migrate" )
return nil
}
isFirstLegacy , firstIdx , err := dbHasLegacyReceipts ( db , 0 )
if err != nil {
return err
}
if ! isFirstLegacy {
log . Info ( "No legacy receipts to migrate" )
return nil
}
log . Info ( "Starting migration" , "ancients" , numAncients , "firstLegacy" , firstIdx )
start := time . Now ( )
if err := db . MigrateTable ( "receipts" , types . ConvertLegacyStoredReceipts ) ; err != nil {
return err
}
if err := db . Close ( ) ; err != nil {
return err
}
log . Info ( "Migration finished" , "duration" , time . Since ( start ) )
return nil
}
// dbHasLegacyReceipts checks freezer entries for legacy receipts. It stops at the first
// non-empty receipt and checks its format. The index of this first non-empty element is
// the second return parameter.
func dbHasLegacyReceipts ( db ethdb . Database , firstIdx uint64 ) ( bool , uint64 , error ) {
// Check first block for legacy receipt format
numAncients , err := db . Ancients ( )
if err != nil {
return false , 0 , err
}
if numAncients < 1 {
return false , 0 , nil
}
if firstIdx >= numAncients {
return false , firstIdx , nil
}
var (
legacy bool
blob [ ] byte
emptyRLPList = [ ] byte { 192 }
)
// Find first block with non-empty receipt, only if
// the index is not already provided.
if firstIdx == 0 {
for i := uint64 ( 0 ) ; i < numAncients ; i ++ {
blob , err = db . Ancient ( "receipts" , i )
if err != nil {
return false , 0 , err
}
if len ( blob ) == 0 {
continue
}
if ! bytes . Equal ( blob , emptyRLPList ) {
firstIdx = i
break
}
}
}
// Is first non-empty receipt legacy?
first , err := db . Ancient ( "receipts" , firstIdx )
if err != nil {
return false , 0 , err
}
legacy , err = types . IsLegacyStoredReceipts ( first )
return legacy , firstIdx , err
}