mirror of https://github.com/ethereum/go-ethereum
Merge pull request #793 from karalabe/discovery-node-cache
p2p/discover: persistent node databasepull/821/head
commit
91cb8cdd2a
@ -0,0 +1,304 @@ |
||||
// Contains the node database, storing previously seen nodes and any collected
|
||||
// metadata about them for QoS purposes.
|
||||
|
||||
package discover |
||||
|
||||
import ( |
||||
"bytes" |
||||
"encoding/binary" |
||||
"os" |
||||
"sync" |
||||
"time" |
||||
|
||||
"github.com/ethereum/go-ethereum/logger" |
||||
"github.com/ethereum/go-ethereum/logger/glog" |
||||
"github.com/ethereum/go-ethereum/rlp" |
||||
"github.com/syndtr/goleveldb/leveldb" |
||||
"github.com/syndtr/goleveldb/leveldb/iterator" |
||||
"github.com/syndtr/goleveldb/leveldb/storage" |
||||
) |
||||
|
||||
var ( |
||||
nodeDBNilNodeID = NodeID{} // Special node ID to use as a nil element.
|
||||
nodeDBNodeExpiration = 24 * time.Hour // Time after which an unseen node should be dropped.
|
||||
nodeDBCleanupCycle = time.Hour // Time period for running the expiration task.
|
||||
) |
||||
|
||||
// nodeDB stores all nodes we know about.
|
||||
type nodeDB struct { |
||||
lvl *leveldb.DB // Interface to the database itself
|
||||
seeder iterator.Iterator // Iterator for fetching possible seed nodes
|
||||
|
||||
runner sync.Once // Ensures we can start at most one expirer
|
||||
quit chan struct{} // Channel to signal the expiring thread to stop
|
||||
} |
||||
|
||||
// Schema layout for the node database
|
||||
var ( |
||||
nodeDBVersionKey = []byte("version") // Version of the database to flush if changes
|
||||
nodeDBItemPrefix = []byte("n:") // Identifier to prefix node entries with
|
||||
|
||||
nodeDBDiscoverRoot = ":discover" |
||||
nodeDBDiscoverPing = nodeDBDiscoverRoot + ":lastping" |
||||
nodeDBDiscoverPong = nodeDBDiscoverRoot + ":lastpong" |
||||
) |
||||
|
||||
// newNodeDB creates a new node database for storing and retrieving infos about
|
||||
// known peers in the network. If no path is given, an in-memory, temporary
|
||||
// database is constructed.
|
||||
func newNodeDB(path string, version int) (*nodeDB, error) { |
||||
if path == "" { |
||||
return newMemoryNodeDB() |
||||
} |
||||
return newPersistentNodeDB(path, version) |
||||
} |
||||
|
||||
// newMemoryNodeDB creates a new in-memory node database without a persistent
|
||||
// backend.
|
||||
func newMemoryNodeDB() (*nodeDB, error) { |
||||
db, err := leveldb.Open(storage.NewMemStorage(), nil) |
||||
if err != nil { |
||||
return nil, err |
||||
} |
||||
return &nodeDB{ |
||||
lvl: db, |
||||
quit: make(chan struct{}), |
||||
}, nil |
||||
} |
||||
|
||||
// newPersistentNodeDB creates/opens a leveldb backed persistent node database,
|
||||
// also flushing its contents in case of a version mismatch.
|
||||
func newPersistentNodeDB(path string, version int) (*nodeDB, error) { |
||||
// Try to open the cache, recovering any corruption
|
||||
db, err := leveldb.OpenFile(path, nil) |
||||
if _, iscorrupted := err.(leveldb.ErrCorrupted); iscorrupted { |
||||
db, err = leveldb.RecoverFile(path, nil) |
||||
} |
||||
if err != nil { |
||||
return nil, err |
||||
} |
||||
// The nodes contained in the cache correspond to a certain protocol version.
|
||||
// Flush all nodes if the version doesn't match.
|
||||
currentVer := make([]byte, binary.MaxVarintLen64) |
||||
currentVer = currentVer[:binary.PutVarint(currentVer, int64(version))] |
||||
|
||||
blob, err := db.Get(nodeDBVersionKey, nil) |
||||
switch err { |
||||
case leveldb.ErrNotFound: |
||||
// Version not found (i.e. empty cache), insert it
|
||||
if err := db.Put(nodeDBVersionKey, currentVer, nil); err != nil { |
||||
db.Close() |
||||
return nil, err |
||||
} |
||||
|
||||
case nil: |
||||
// Version present, flush if different
|
||||
if !bytes.Equal(blob, currentVer) { |
||||
db.Close() |
||||
if err = os.RemoveAll(path); err != nil { |
||||
return nil, err |
||||
} |
||||
return newPersistentNodeDB(path, version) |
||||
} |
||||
} |
||||
return &nodeDB{ |
||||
lvl: db, |
||||
quit: make(chan struct{}), |
||||
}, nil |
||||
} |
||||
|
||||
// makeKey generates the leveldb key-blob from a node id and its particular
|
||||
// field of interest.
|
||||
func makeKey(id NodeID, field string) []byte { |
||||
if bytes.Equal(id[:], nodeDBNilNodeID[:]) { |
||||
return []byte(field) |
||||
} |
||||
return append(nodeDBItemPrefix, append(id[:], field...)...) |
||||
} |
||||
|
||||
// splitKey tries to split a database key into a node id and a field part.
|
||||
func splitKey(key []byte) (id NodeID, field string) { |
||||
// If the key is not of a node, return it plainly
|
||||
if !bytes.HasPrefix(key, nodeDBItemPrefix) { |
||||
return NodeID{}, string(key) |
||||
} |
||||
// Otherwise split the id and field
|
||||
item := key[len(nodeDBItemPrefix):] |
||||
copy(id[:], item[:len(id)]) |
||||
field = string(item[len(id):]) |
||||
|
||||
return id, field |
||||
} |
||||
|
||||
// fetchInt64 retrieves an integer instance associated with a particular
|
||||
// database key.
|
||||
func (db *nodeDB) fetchInt64(key []byte) int64 { |
||||
blob, err := db.lvl.Get(key, nil) |
||||
if err != nil { |
||||
return 0 |
||||
} |
||||
val, read := binary.Varint(blob) |
||||
if read <= 0 { |
||||
return 0 |
||||
} |
||||
return val |
||||
} |
||||
|
||||
// storeInt64 update a specific database entry to the current time instance as a
|
||||
// unix timestamp.
|
||||
func (db *nodeDB) storeInt64(key []byte, n int64) error { |
||||
blob := make([]byte, binary.MaxVarintLen64) |
||||
blob = blob[:binary.PutVarint(blob, n)] |
||||
|
||||
return db.lvl.Put(key, blob, nil) |
||||
} |
||||
|
||||
// node retrieves a node with a given id from the database.
|
||||
func (db *nodeDB) node(id NodeID) *Node { |
||||
blob, err := db.lvl.Get(makeKey(id, nodeDBDiscoverRoot), nil) |
||||
if err != nil { |
||||
glog.V(logger.Detail).Infof("failed to retrieve node %v: %v", id, err) |
||||
return nil |
||||
} |
||||
node := new(Node) |
||||
if err := rlp.DecodeBytes(blob, node); err != nil { |
||||
glog.V(logger.Warn).Infof("failed to decode node RLP: %v", err) |
||||
return nil |
||||
} |
||||
return node |
||||
} |
||||
|
||||
// updateNode inserts - potentially overwriting - a node into the peer database.
|
||||
func (db *nodeDB) updateNode(node *Node) error { |
||||
blob, err := rlp.EncodeToBytes(node) |
||||
if err != nil { |
||||
return err |
||||
} |
||||
return db.lvl.Put(makeKey(node.ID, nodeDBDiscoverRoot), blob, nil) |
||||
} |
||||
|
||||
// ensureExpirer is a small helper method ensuring that the data expiration
|
||||
// mechanism is running. If the expiration goroutine is already running, this
|
||||
// method simply returns.
|
||||
//
|
||||
// The goal is to start the data evacuation only after the network successfully
|
||||
// bootstrapped itself (to prevent dumping potentially useful seed nodes). Since
|
||||
// it would require significant overhead to exactly trace the first successful
|
||||
// convergence, it's simpler to "ensure" the correct state when an appropriate
|
||||
// condition occurs (i.e. a successful bonding), and discard further events.
|
||||
func (db *nodeDB) ensureExpirer() { |
||||
db.runner.Do(func() { go db.expirer() }) |
||||
} |
||||
|
||||
// expirer should be started in a go routine, and is responsible for looping ad
|
||||
// infinitum and dropping stale data from the database.
|
||||
func (db *nodeDB) expirer() { |
||||
tick := time.Tick(nodeDBCleanupCycle) |
||||
for { |
||||
select { |
||||
case <-tick: |
||||
if err := db.expireNodes(); err != nil { |
||||
glog.V(logger.Error).Infof("Failed to expire nodedb items: %v", err) |
||||
} |
||||
|
||||
case <-db.quit: |
||||
return |
||||
} |
||||
} |
||||
} |
||||
|
||||
// expireNodes iterates over the database and deletes all nodes that have not
|
||||
// been seen (i.e. received a pong from) for some alloted time.
|
||||
func (db *nodeDB) expireNodes() error { |
||||
threshold := time.Now().Add(-nodeDBNodeExpiration) |
||||
|
||||
// Find discovered nodes that are older than the allowance
|
||||
it := db.lvl.NewIterator(nil, nil) |
||||
defer it.Release() |
||||
|
||||
for it.Next() { |
||||
// Skip the item if not a discovery node
|
||||
id, field := splitKey(it.Key()) |
||||
if field != nodeDBDiscoverRoot { |
||||
continue |
||||
} |
||||
// Skip the node if not expired yet
|
||||
if seen := db.lastPong(id); seen.After(threshold) { |
||||
continue |
||||
} |
||||
// Otherwise delete all associated information
|
||||
prefix := makeKey(id, "") |
||||
for ok := it.Seek(prefix); ok && bytes.HasPrefix(it.Key(), prefix); ok = it.Next() { |
||||
if err := db.lvl.Delete(it.Key(), nil); err != nil { |
||||
return err |
||||
} |
||||
} |
||||
} |
||||
return nil |
||||
} |
||||
|
||||
// lastPing retrieves the time of the last ping packet send to a remote node,
|
||||
// requesting binding.
|
||||
func (db *nodeDB) lastPing(id NodeID) time.Time { |
||||
return time.Unix(db.fetchInt64(makeKey(id, nodeDBDiscoverPing)), 0) |
||||
} |
||||
|
||||
// updateLastPing updates the last time we tried contacting a remote node.
|
||||
func (db *nodeDB) updateLastPing(id NodeID, instance time.Time) error { |
||||
return db.storeInt64(makeKey(id, nodeDBDiscoverPing), instance.Unix()) |
||||
} |
||||
|
||||
// lastPong retrieves the time of the last successful contact from remote node.
|
||||
func (db *nodeDB) lastPong(id NodeID) time.Time { |
||||
return time.Unix(db.fetchInt64(makeKey(id, nodeDBDiscoverPong)), 0) |
||||
} |
||||
|
||||
// updateLastPong updates the last time a remote node successfully contacted.
|
||||
func (db *nodeDB) updateLastPong(id NodeID, instance time.Time) error { |
||||
return db.storeInt64(makeKey(id, nodeDBDiscoverPong), instance.Unix()) |
||||
} |
||||
|
||||
// querySeeds retrieves a batch of nodes to be used as potential seed servers
|
||||
// during bootstrapping the node into the network.
|
||||
//
|
||||
// Ideal seeds are the most recently seen nodes (highest probability to be still
|
||||
// alive), but yet untried. However, since leveldb only supports dumb iteration
|
||||
// we will instead start pulling in potential seeds that haven't been yet pinged
|
||||
// since the start of the boot procedure.
|
||||
//
|
||||
// If the database runs out of potential seeds, we restart the startup counter
|
||||
// and start iterating over the peers again.
|
||||
func (db *nodeDB) querySeeds(n int) []*Node { |
||||
// Create a new seed iterator if none exists
|
||||
if db.seeder == nil { |
||||
db.seeder = db.lvl.NewIterator(nil, nil) |
||||
} |
||||
// Iterate over the nodes and find suitable seeds
|
||||
nodes := make([]*Node, 0, n) |
||||
for len(nodes) < n && db.seeder.Next() { |
||||
// Iterate until a discovery node is found
|
||||
id, field := splitKey(db.seeder.Key()) |
||||
if field != nodeDBDiscoverRoot { |
||||
continue |
||||
} |
||||
// Load it as a potential seed
|
||||
if node := db.node(id); node != nil { |
||||
nodes = append(nodes, node) |
||||
} |
||||
} |
||||
// Release the iterator if we reached the end
|
||||
if len(nodes) == 0 { |
||||
db.seeder.Release() |
||||
db.seeder = nil |
||||
} |
||||
return nodes |
||||
} |
||||
|
||||
// close flushes and closes the database files.
|
||||
func (db *nodeDB) close() { |
||||
if db.seeder != nil { |
||||
db.seeder.Release() |
||||
} |
||||
close(db.quit) |
||||
db.lvl.Close() |
||||
} |
@ -0,0 +1,313 @@ |
||||
package discover |
||||
|
||||
import ( |
||||
"bytes" |
||||
"io/ioutil" |
||||
"net" |
||||
"os" |
||||
"path/filepath" |
||||
"testing" |
||||
"time" |
||||
) |
||||
|
||||
var nodeDBKeyTests = []struct { |
||||
id NodeID |
||||
field string |
||||
key []byte |
||||
}{ |
||||
{ |
||||
id: NodeID{}, |
||||
field: "version", |
||||
key: []byte{0x76, 0x65, 0x72, 0x73, 0x69, 0x6f, 0x6e}, // field
|
||||
}, |
||||
{ |
||||
id: MustHexID("0x1dd9d65c4552b5eb43d5ad55a2ee3f56c6cbc1c64a5c8d659f51fcd51bace24351232b8d7821617d2b29b54b81cdefb9b3e9c37d7fd5f63270bcc9e1a6f6a439"), |
||||
field: ":discover", |
||||
key: []byte{0x6e, 0x3a, // prefix
|
||||
0x1d, 0xd9, 0xd6, 0x5c, 0x45, 0x52, 0xb5, 0xeb, // node id
|
||||
0x43, 0xd5, 0xad, 0x55, 0xa2, 0xee, 0x3f, 0x56, //
|
||||
0xc6, 0xcb, 0xc1, 0xc6, 0x4a, 0x5c, 0x8d, 0x65, //
|
||||
0x9f, 0x51, 0xfc, 0xd5, 0x1b, 0xac, 0xe2, 0x43, //
|
||||
0x51, 0x23, 0x2b, 0x8d, 0x78, 0x21, 0x61, 0x7d, //
|
||||
0x2b, 0x29, 0xb5, 0x4b, 0x81, 0xcd, 0xef, 0xb9, //
|
||||
0xb3, 0xe9, 0xc3, 0x7d, 0x7f, 0xd5, 0xf6, 0x32, //
|
||||
0x70, 0xbc, 0xc9, 0xe1, 0xa6, 0xf6, 0xa4, 0x39, //
|
||||
0x3a, 0x64, 0x69, 0x73, 0x63, 0x6f, 0x76, 0x65, 0x72, // field
|
||||
}, |
||||
}, |
||||
} |
||||
|
||||
func TestNodeDBKeys(t *testing.T) { |
||||
for i, tt := range nodeDBKeyTests { |
||||
if key := makeKey(tt.id, tt.field); !bytes.Equal(key, tt.key) { |
||||
t.Errorf("make test %d: key mismatch: have 0x%x, want 0x%x", i, key, tt.key) |
||||
} |
||||
id, field := splitKey(tt.key) |
||||
if !bytes.Equal(id[:], tt.id[:]) { |
||||
t.Errorf("split test %d: id mismatch: have 0x%x, want 0x%x", i, id, tt.id) |
||||
} |
||||
if field != tt.field { |
||||
t.Errorf("split test %d: field mismatch: have 0x%x, want 0x%x", i, field, tt.field) |
||||
} |
||||
} |
||||
} |
||||
|
||||
var nodeDBInt64Tests = []struct { |
||||
key []byte |
||||
value int64 |
||||
}{ |
||||
{key: []byte{0x01}, value: 1}, |
||||
{key: []byte{0x02}, value: 2}, |
||||
{key: []byte{0x03}, value: 3}, |
||||
} |
||||
|
||||
func TestNodeDBInt64(t *testing.T) { |
||||
db, _ := newNodeDB("", Version) |
||||
defer db.close() |
||||
|
||||
tests := nodeDBInt64Tests |
||||
for i := 0; i < len(tests); i++ { |
||||
// Insert the next value
|
||||
if err := db.storeInt64(tests[i].key, tests[i].value); err != nil { |
||||
t.Errorf("test %d: failed to store value: %v", i, err) |
||||
} |
||||
// Check all existing and non existing values
|
||||
for j := 0; j < len(tests); j++ { |
||||
num := db.fetchInt64(tests[j].key) |
||||
switch { |
||||
case j <= i && num != tests[j].value: |
||||
t.Errorf("test %d, item %d: value mismatch: have %v, want %v", i, j, num, tests[j].value) |
||||
case j > i && num != 0: |
||||
t.Errorf("test %d, item %d: value mismatch: have %v, want %v", i, j, num, 0) |
||||
} |
||||
} |
||||
} |
||||
} |
||||
|
||||
func TestNodeDBFetchStore(t *testing.T) { |
||||
node := &Node{ |
||||
ID: MustHexID("0x1dd9d65c4552b5eb43d5ad55a2ee3f56c6cbc1c64a5c8d659f51fcd51bace24351232b8d7821617d2b29b54b81cdefb9b3e9c37d7fd5f63270bcc9e1a6f6a439"), |
||||
IP: net.IP([]byte{192, 168, 0, 1}), |
||||
TCPPort: 30303, |
||||
} |
||||
inst := time.Now() |
||||
|
||||
db, _ := newNodeDB("", Version) |
||||
defer db.close() |
||||
|
||||
// Check fetch/store operations on a node ping object
|
||||
if stored := db.lastPing(node.ID); stored.Unix() != 0 { |
||||
t.Errorf("ping: non-existing object: %v", stored) |
||||
} |
||||
if err := db.updateLastPing(node.ID, inst); err != nil { |
||||
t.Errorf("ping: failed to update: %v", err) |
||||
} |
||||
if stored := db.lastPing(node.ID); stored.Unix() != inst.Unix() { |
||||
t.Errorf("ping: value mismatch: have %v, want %v", stored, inst) |
||||
} |
||||
// Check fetch/store operations on a node pong object
|
||||
if stored := db.lastPong(node.ID); stored.Unix() != 0 { |
||||
t.Errorf("pong: non-existing object: %v", stored) |
||||
} |
||||
if err := db.updateLastPong(node.ID, inst); err != nil { |
||||
t.Errorf("pong: failed to update: %v", err) |
||||
} |
||||
if stored := db.lastPong(node.ID); stored.Unix() != inst.Unix() { |
||||
t.Errorf("pong: value mismatch: have %v, want %v", stored, inst) |
||||
} |
||||
// Check fetch/store operations on an actual node object
|
||||
if stored := db.node(node.ID); stored != nil { |
||||
t.Errorf("node: non-existing object: %v", stored) |
||||
} |
||||
if err := db.updateNode(node); err != nil { |
||||
t.Errorf("node: failed to update: %v", err) |
||||
} |
||||
if stored := db.node(node.ID); stored == nil { |
||||
t.Errorf("node: not found") |
||||
} else if !bytes.Equal(stored.ID[:], node.ID[:]) || !stored.IP.Equal(node.IP) || stored.TCPPort != node.TCPPort { |
||||
t.Errorf("node: data mismatch: have %v, want %v", stored, node) |
||||
} |
||||
} |
||||
|
||||
var nodeDBSeedQueryNodes = []struct { |
||||
node Node |
||||
pong time.Time |
||||
}{ |
||||
{ |
||||
node: Node{ |
||||
ID: MustHexID("0x01d9d65c4552b5eb43d5ad55a2ee3f56c6cbc1c64a5c8d659f51fcd51bace24351232b8d7821617d2b29b54b81cdefb9b3e9c37d7fd5f63270bcc9e1a6f6a439"), |
||||
IP: []byte{127, 0, 0, 1}, |
||||
}, |
||||
pong: time.Now().Add(-2 * time.Second), |
||||
}, |
||||
{ |
||||
node: Node{ |
||||
ID: MustHexID("0x02d9d65c4552b5eb43d5ad55a2ee3f56c6cbc1c64a5c8d659f51fcd51bace24351232b8d7821617d2b29b54b81cdefb9b3e9c37d7fd5f63270bcc9e1a6f6a439"), |
||||
IP: []byte{127, 0, 0, 2}, |
||||
}, |
||||
pong: time.Now().Add(-3 * time.Second), |
||||
}, |
||||
{ |
||||
node: Node{ |
||||
ID: MustHexID("0x03d9d65c4552b5eb43d5ad55a2ee3f56c6cbc1c64a5c8d659f51fcd51bace24351232b8d7821617d2b29b54b81cdefb9b3e9c37d7fd5f63270bcc9e1a6f6a439"), |
||||
IP: []byte{127, 0, 0, 3}, |
||||
}, |
||||
pong: time.Now().Add(-1 * time.Second), |
||||
}, |
||||
} |
||||
|
||||
func TestNodeDBSeedQuery(t *testing.T) { |
||||
db, _ := newNodeDB("", Version) |
||||
defer db.close() |
||||
|
||||
// Insert a batch of nodes for querying
|
||||
for i, seed := range nodeDBSeedQueryNodes { |
||||
if err := db.updateNode(&seed.node); err != nil { |
||||
t.Fatalf("node %d: failed to insert: %v", i, err) |
||||
} |
||||
} |
||||
// Retrieve the entire batch and check for duplicates
|
||||
seeds := db.querySeeds(2 * len(nodeDBSeedQueryNodes)) |
||||
if len(seeds) != len(nodeDBSeedQueryNodes) { |
||||
t.Errorf("seed count mismatch: have %v, want %v", len(seeds), len(nodeDBSeedQueryNodes)) |
||||
} |
||||
have := make(map[NodeID]struct{}) |
||||
for _, seed := range seeds { |
||||
have[seed.ID] = struct{}{} |
||||
} |
||||
want := make(map[NodeID]struct{}) |
||||
for _, seed := range nodeDBSeedQueryNodes { |
||||
want[seed.node.ID] = struct{}{} |
||||
} |
||||
for id, _ := range have { |
||||
if _, ok := want[id]; !ok { |
||||
t.Errorf("extra seed: %v", id) |
||||
} |
||||
} |
||||
for id, _ := range want { |
||||
if _, ok := have[id]; !ok { |
||||
t.Errorf("missing seed: %v", id) |
||||
} |
||||
} |
||||
// Make sure the next batch is empty (seed EOF)
|
||||
seeds = db.querySeeds(2 * len(nodeDBSeedQueryNodes)) |
||||
if len(seeds) != 0 { |
||||
t.Errorf("seed count mismatch: have %v, want %v", len(seeds), 0) |
||||
} |
||||
} |
||||
|
||||
func TestNodeDBSeedQueryContinuation(t *testing.T) { |
||||
db, _ := newNodeDB("", Version) |
||||
defer db.close() |
||||
|
||||
// Insert a batch of nodes for querying
|
||||
for i, seed := range nodeDBSeedQueryNodes { |
||||
if err := db.updateNode(&seed.node); err != nil { |
||||
t.Fatalf("node %d: failed to insert: %v", i, err) |
||||
} |
||||
} |
||||
// Iteratively retrieve the batch, checking for an empty batch on reset
|
||||
for i := 0; i < len(nodeDBSeedQueryNodes); i++ { |
||||
if seeds := db.querySeeds(1); len(seeds) != 1 { |
||||
t.Errorf("1st iteration %d: seed count mismatch: have %v, want %v", i, len(seeds), 1) |
||||
} |
||||
} |
||||
if seeds := db.querySeeds(1); len(seeds) != 0 { |
||||
t.Errorf("reset: seed count mismatch: have %v, want %v", len(seeds), 0) |
||||
} |
||||
for i := 0; i < len(nodeDBSeedQueryNodes); i++ { |
||||
if seeds := db.querySeeds(1); len(seeds) != 1 { |
||||
t.Errorf("2nd iteration %d: seed count mismatch: have %v, want %v", i, len(seeds), 1) |
||||
} |
||||
} |
||||
} |
||||
|
||||
func TestNodeDBPersistency(t *testing.T) { |
||||
root, err := ioutil.TempDir("", "nodedb-") |
||||
if err != nil { |
||||
t.Fatalf("failed to create temporary data folder: %v", err) |
||||
} |
||||
defer os.RemoveAll(root) |
||||
|
||||
var ( |
||||
testKey = []byte("somekey") |
||||
testInt = int64(314) |
||||
) |
||||
|
||||
// Create a persistent database and store some values
|
||||
db, err := newNodeDB(filepath.Join("root", "database"), Version) |
||||
if err != nil { |
||||
t.Fatalf("failed to create persistent database: %v", err) |
||||
} |
||||
if err := db.storeInt64(testKey, testInt); err != nil { |
||||
t.Fatalf("failed to store value: %v.", err) |
||||
} |
||||
db.close() |
||||
|
||||
// Reopen the database and check the value
|
||||
db, err = newNodeDB(filepath.Join("root", "database"), Version) |
||||
if err != nil { |
||||
t.Fatalf("failed to open persistent database: %v", err) |
||||
} |
||||
if val := db.fetchInt64(testKey); val != testInt { |
||||
t.Fatalf("value mismatch: have %v, want %v", val, testInt) |
||||
} |
||||
db.close() |
||||
|
||||
// Change the database version and check flush
|
||||
db, err = newNodeDB(filepath.Join("root", "database"), Version+1) |
||||
if err != nil { |
||||
t.Fatalf("failed to open persistent database: %v", err) |
||||
} |
||||
if val := db.fetchInt64(testKey); val != 0 { |
||||
t.Fatalf("value mismatch: have %v, want %v", val, 0) |
||||
} |
||||
db.close() |
||||
} |
||||
|
||||
var nodeDBExpirationNodes = []struct { |
||||
node Node |
||||
pong time.Time |
||||
exp bool |
||||
}{ |
||||
{ |
||||
node: Node{ |
||||
ID: MustHexID("0x01d9d65c4552b5eb43d5ad55a2ee3f56c6cbc1c64a5c8d659f51fcd51bace24351232b8d7821617d2b29b54b81cdefb9b3e9c37d7fd5f63270bcc9e1a6f6a439"), |
||||
IP: []byte{127, 0, 0, 1}, |
||||
}, |
||||
pong: time.Now().Add(-nodeDBNodeExpiration + time.Minute), |
||||
exp: false, |
||||
}, { |
||||
node: Node{ |
||||
ID: MustHexID("0x02d9d65c4552b5eb43d5ad55a2ee3f56c6cbc1c64a5c8d659f51fcd51bace24351232b8d7821617d2b29b54b81cdefb9b3e9c37d7fd5f63270bcc9e1a6f6a439"), |
||||
IP: []byte{127, 0, 0, 2}, |
||||
}, |
||||
pong: time.Now().Add(-nodeDBNodeExpiration - time.Minute), |
||||
exp: true, |
||||
}, |
||||
} |
||||
|
||||
func TestNodeDBExpiration(t *testing.T) { |
||||
db, _ := newNodeDB("", Version) |
||||
defer db.close() |
||||
|
||||
// Add all the test nodes and set their last pong time
|
||||
for i, seed := range nodeDBExpirationNodes { |
||||
if err := db.updateNode(&seed.node); err != nil { |
||||
t.Fatalf("node %d: failed to insert: %v", i, err) |
||||
} |
||||
if err := db.updateLastPong(seed.node.ID, seed.pong); err != nil { |
||||
t.Fatalf("node %d: failed to update pong: %v", i, err) |
||||
} |
||||
} |
||||
// Expire some of them, and check the rest
|
||||
if err := db.expireNodes(); err != nil { |
||||
t.Fatalf("failed to expire nodes: %v", err) |
||||
} |
||||
for i, seed := range nodeDBExpirationNodes { |
||||
node := db.node(seed.node.ID) |
||||
if (node == nil && !seed.exp) || (node != nil && seed.exp) { |
||||
t.Errorf("node %d: expiration mismatch: have %v, want %v", i, node, seed.exp) |
||||
} |
||||
} |
||||
} |
Loading…
Reference in new issue