mirror of
https://github.com/lightningnetwork/lnd.git
synced 2024-11-19 09:53:54 +01:00
97c73706b5
We use the event timestamp of a forwarding event as its primary storage key. On systems with a bad clock resolution this can lead to collisions of the events if some of the timestamps are identical. We fix this problem by shifting the timestamps on the nanosecond level until only unique values remain.
384 lines
12 KiB
Go
384 lines
12 KiB
Go
package channeldb
|
|
|
|
import (
|
|
"math/rand"
|
|
"reflect"
|
|
"testing"
|
|
"time"
|
|
|
|
"github.com/davecgh/go-spew/spew"
|
|
"github.com/lightningnetwork/lnd/lnwire"
|
|
"github.com/stretchr/testify/assert"
|
|
)
|
|
|
|
// TestForwardingLogBasicStorageAndQuery tests that we're able to store and
|
|
// then query for items that have previously been added to the event log.
|
|
func TestForwardingLogBasicStorageAndQuery(t *testing.T) {
|
|
t.Parallel()
|
|
|
|
// First, we'll set up a test database, and use that to instantiate the
|
|
// forwarding event log that we'll be using for the duration of the
|
|
// test.
|
|
db, cleanUp, err := MakeTestDB()
|
|
if err != nil {
|
|
t.Fatalf("unable to make test db: %v", err)
|
|
}
|
|
defer cleanUp()
|
|
|
|
log := ForwardingLog{
|
|
db: db,
|
|
}
|
|
|
|
initialTime := time.Unix(1234, 0)
|
|
timestamp := time.Unix(1234, 0)
|
|
|
|
// We'll create 100 random events, which each event being spaced 10
|
|
// minutes after the prior event.
|
|
numEvents := 100
|
|
events := make([]ForwardingEvent, numEvents)
|
|
for i := 0; i < numEvents; i++ {
|
|
events[i] = ForwardingEvent{
|
|
Timestamp: timestamp,
|
|
IncomingChanID: lnwire.NewShortChanIDFromInt(uint64(rand.Int63())),
|
|
OutgoingChanID: lnwire.NewShortChanIDFromInt(uint64(rand.Int63())),
|
|
AmtIn: lnwire.MilliSatoshi(rand.Int63()),
|
|
AmtOut: lnwire.MilliSatoshi(rand.Int63()),
|
|
}
|
|
|
|
timestamp = timestamp.Add(time.Minute * 10)
|
|
}
|
|
|
|
// Now that all of our set of events constructed, we'll add them to the
|
|
// database in a batch manner.
|
|
if err := log.AddForwardingEvents(events); err != nil {
|
|
t.Fatalf("unable to add events: %v", err)
|
|
}
|
|
|
|
// With our events added we'll now construct a basic query to retrieve
|
|
// all of the events.
|
|
eventQuery := ForwardingEventQuery{
|
|
StartTime: initialTime,
|
|
EndTime: timestamp,
|
|
IndexOffset: 0,
|
|
NumMaxEvents: 1000,
|
|
}
|
|
timeSlice, err := log.Query(eventQuery)
|
|
if err != nil {
|
|
t.Fatalf("unable to query for events: %v", err)
|
|
}
|
|
|
|
// The set of returned events should match identically, as they should
|
|
// be returned in sorted order.
|
|
if !reflect.DeepEqual(events, timeSlice.ForwardingEvents) {
|
|
t.Fatalf("event mismatch: expected %v vs %v",
|
|
spew.Sdump(events), spew.Sdump(timeSlice.ForwardingEvents))
|
|
}
|
|
|
|
// The offset index of the final entry should be numEvents, so the
|
|
// number of total events we've written.
|
|
if timeSlice.LastIndexOffset != uint32(numEvents) {
|
|
t.Fatalf("wrong final offset: expected %v, got %v",
|
|
timeSlice.LastIndexOffset, numEvents)
|
|
}
|
|
}
|
|
|
|
// TestForwardingLogQueryOptions tests that the query offset works properly. So
|
|
// if we add a series of events, then we should be able to seek within the
|
|
// timeslice accordingly. This exercises the index offset and num max event
|
|
// field in the query, and also the last index offset field int he response.
|
|
func TestForwardingLogQueryOptions(t *testing.T) {
|
|
t.Parallel()
|
|
|
|
// First, we'll set up a test database, and use that to instantiate the
|
|
// forwarding event log that we'll be using for the duration of the
|
|
// test.
|
|
db, cleanUp, err := MakeTestDB()
|
|
if err != nil {
|
|
t.Fatalf("unable to make test db: %v", err)
|
|
}
|
|
defer cleanUp()
|
|
|
|
log := ForwardingLog{
|
|
db: db,
|
|
}
|
|
|
|
initialTime := time.Unix(1234, 0)
|
|
endTime := time.Unix(1234, 0)
|
|
|
|
// We'll create 20 random events, which each event being spaced 10
|
|
// minutes after the prior event.
|
|
numEvents := 20
|
|
events := make([]ForwardingEvent, numEvents)
|
|
for i := 0; i < numEvents; i++ {
|
|
events[i] = ForwardingEvent{
|
|
Timestamp: endTime,
|
|
IncomingChanID: lnwire.NewShortChanIDFromInt(uint64(rand.Int63())),
|
|
OutgoingChanID: lnwire.NewShortChanIDFromInt(uint64(rand.Int63())),
|
|
AmtIn: lnwire.MilliSatoshi(rand.Int63()),
|
|
AmtOut: lnwire.MilliSatoshi(rand.Int63()),
|
|
}
|
|
|
|
endTime = endTime.Add(time.Minute * 10)
|
|
}
|
|
|
|
// Now that all of our set of events constructed, we'll add them to the
|
|
// database in a batch manner.
|
|
if err := log.AddForwardingEvents(events); err != nil {
|
|
t.Fatalf("unable to add events: %v", err)
|
|
}
|
|
|
|
// With all of our events added, we should be able to query for the
|
|
// first 10 events using the max event query field.
|
|
eventQuery := ForwardingEventQuery{
|
|
StartTime: initialTime,
|
|
EndTime: endTime,
|
|
IndexOffset: 0,
|
|
NumMaxEvents: 10,
|
|
}
|
|
timeSlice, err := log.Query(eventQuery)
|
|
if err != nil {
|
|
t.Fatalf("unable to query for events: %v", err)
|
|
}
|
|
|
|
// We should get exactly 10 events back.
|
|
if len(timeSlice.ForwardingEvents) != 10 {
|
|
t.Fatalf("wrong number of events: expected %v, got %v", 10,
|
|
len(timeSlice.ForwardingEvents))
|
|
}
|
|
|
|
// The set of events returned should be the first 10 events that we
|
|
// added.
|
|
if !reflect.DeepEqual(events[:10], timeSlice.ForwardingEvents) {
|
|
t.Fatalf("wrong response: expected %v, got %v",
|
|
spew.Sdump(events[:10]),
|
|
spew.Sdump(timeSlice.ForwardingEvents))
|
|
}
|
|
|
|
// The final offset should be the exact number of events returned.
|
|
if timeSlice.LastIndexOffset != 10 {
|
|
t.Fatalf("wrong index offset: expected %v, got %v", 10,
|
|
timeSlice.LastIndexOffset)
|
|
}
|
|
|
|
// If we use the final offset to query again, then we should get 10
|
|
// more events, that are the last 10 events we wrote.
|
|
eventQuery.IndexOffset = 10
|
|
timeSlice, err = log.Query(eventQuery)
|
|
if err != nil {
|
|
t.Fatalf("unable to query for events: %v", err)
|
|
}
|
|
|
|
// We should get exactly 10 events back once again.
|
|
if len(timeSlice.ForwardingEvents) != 10 {
|
|
t.Fatalf("wrong number of events: expected %v, got %v", 10,
|
|
len(timeSlice.ForwardingEvents))
|
|
}
|
|
|
|
// The events that we got back should be the last 10 events that we
|
|
// wrote out.
|
|
if !reflect.DeepEqual(events[10:], timeSlice.ForwardingEvents) {
|
|
t.Fatalf("wrong response: expected %v, got %v",
|
|
spew.Sdump(events[10:]),
|
|
spew.Sdump(timeSlice.ForwardingEvents))
|
|
}
|
|
|
|
// Finally, the last index offset should be 20, or the number of
|
|
// records we've written out.
|
|
if timeSlice.LastIndexOffset != 20 {
|
|
t.Fatalf("wrong index offset: expected %v, got %v", 20,
|
|
timeSlice.LastIndexOffset)
|
|
}
|
|
}
|
|
|
|
// TestForwardingLogQueryLimit tests that we're able to properly limit the
|
|
// number of events that are returned as part of a query.
|
|
func TestForwardingLogQueryLimit(t *testing.T) {
|
|
t.Parallel()
|
|
|
|
// First, we'll set up a test database, and use that to instantiate the
|
|
// forwarding event log that we'll be using for the duration of the
|
|
// test.
|
|
db, cleanUp, err := MakeTestDB()
|
|
if err != nil {
|
|
t.Fatalf("unable to make test db: %v", err)
|
|
}
|
|
defer cleanUp()
|
|
|
|
log := ForwardingLog{
|
|
db: db,
|
|
}
|
|
|
|
initialTime := time.Unix(1234, 0)
|
|
endTime := time.Unix(1234, 0)
|
|
|
|
// We'll create 200 random events, which each event being spaced 10
|
|
// minutes after the prior event.
|
|
numEvents := 200
|
|
events := make([]ForwardingEvent, numEvents)
|
|
for i := 0; i < numEvents; i++ {
|
|
events[i] = ForwardingEvent{
|
|
Timestamp: endTime,
|
|
IncomingChanID: lnwire.NewShortChanIDFromInt(uint64(rand.Int63())),
|
|
OutgoingChanID: lnwire.NewShortChanIDFromInt(uint64(rand.Int63())),
|
|
AmtIn: lnwire.MilliSatoshi(rand.Int63()),
|
|
AmtOut: lnwire.MilliSatoshi(rand.Int63()),
|
|
}
|
|
|
|
endTime = endTime.Add(time.Minute * 10)
|
|
}
|
|
|
|
// Now that all of our set of events constructed, we'll add them to the
|
|
// database in a batch manner.
|
|
if err := log.AddForwardingEvents(events); err != nil {
|
|
t.Fatalf("unable to add events: %v", err)
|
|
}
|
|
|
|
// Once the events have been written out, we'll issue a query over the
|
|
// entire range, but restrict the number of events to the first 100.
|
|
eventQuery := ForwardingEventQuery{
|
|
StartTime: initialTime,
|
|
EndTime: endTime,
|
|
IndexOffset: 0,
|
|
NumMaxEvents: 100,
|
|
}
|
|
timeSlice, err := log.Query(eventQuery)
|
|
if err != nil {
|
|
t.Fatalf("unable to query for events: %v", err)
|
|
}
|
|
|
|
// We should get exactly 100 events back.
|
|
if len(timeSlice.ForwardingEvents) != 100 {
|
|
t.Fatalf("wrong number of events: expected %v, got %v", 10,
|
|
len(timeSlice.ForwardingEvents))
|
|
}
|
|
|
|
// The set of events returned should be the first 100 events that we
|
|
// added.
|
|
if !reflect.DeepEqual(events[:100], timeSlice.ForwardingEvents) {
|
|
t.Fatalf("wrong response: expected %v, got %v",
|
|
spew.Sdump(events[:100]),
|
|
spew.Sdump(timeSlice.ForwardingEvents))
|
|
}
|
|
|
|
// The final offset should be the exact number of events returned.
|
|
if timeSlice.LastIndexOffset != 100 {
|
|
t.Fatalf("wrong index offset: expected %v, got %v", 100,
|
|
timeSlice.LastIndexOffset)
|
|
}
|
|
}
|
|
|
|
// TestForwardingLogMakeUniqueTimestamps makes sure the function that creates
|
|
// unique timestamps does it job correctly.
|
|
func TestForwardingLogMakeUniqueTimestamps(t *testing.T) {
|
|
t.Parallel()
|
|
|
|
// Create a list of events where some of the timestamps collide. We
|
|
// expect no existing timestamp to be overwritten, instead the "gaps"
|
|
// between them should be filled.
|
|
inputSlice := []ForwardingEvent{
|
|
{Timestamp: time.Unix(0, 1001)},
|
|
{Timestamp: time.Unix(0, 2001)},
|
|
{Timestamp: time.Unix(0, 1001)},
|
|
{Timestamp: time.Unix(0, 1002)},
|
|
{Timestamp: time.Unix(0, 1004)},
|
|
{Timestamp: time.Unix(0, 1004)},
|
|
{Timestamp: time.Unix(0, 1007)},
|
|
{Timestamp: time.Unix(0, 1001)},
|
|
}
|
|
expectedSlice := []ForwardingEvent{
|
|
{Timestamp: time.Unix(0, 1001)},
|
|
{Timestamp: time.Unix(0, 1002)},
|
|
{Timestamp: time.Unix(0, 1003)},
|
|
{Timestamp: time.Unix(0, 1004)},
|
|
{Timestamp: time.Unix(0, 1005)},
|
|
{Timestamp: time.Unix(0, 1006)},
|
|
{Timestamp: time.Unix(0, 1007)},
|
|
{Timestamp: time.Unix(0, 2001)},
|
|
}
|
|
|
|
makeUniqueTimestamps(inputSlice)
|
|
|
|
for idx, in := range inputSlice {
|
|
expect := expectedSlice[idx]
|
|
assert.Equal(
|
|
t, expect.Timestamp.UnixNano(), in.Timestamp.UnixNano(),
|
|
)
|
|
}
|
|
}
|
|
|
|
// TestForwardingLogStoreEvent makes sure forwarding events are stored without
|
|
// colliding on duplicate timestamps.
|
|
func TestForwardingLogStoreEvent(t *testing.T) {
|
|
t.Parallel()
|
|
|
|
// First, we'll set up a test database, and use that to instantiate the
|
|
// forwarding event log that we'll be using for the duration of the
|
|
// test.
|
|
db, cleanUp, err := MakeTestDB()
|
|
if err != nil {
|
|
t.Fatalf("unable to make test db: %v", err)
|
|
}
|
|
defer cleanUp()
|
|
|
|
log := ForwardingLog{
|
|
db: db,
|
|
}
|
|
|
|
// We'll create 20 random events, with each event having a timestamp
|
|
// with just one nanosecond apart.
|
|
numEvents := 20
|
|
events := make([]ForwardingEvent, numEvents)
|
|
ts := time.Now().UnixNano()
|
|
for i := 0; i < numEvents; i++ {
|
|
events[i] = ForwardingEvent{
|
|
Timestamp: time.Unix(0, ts+int64(i)),
|
|
IncomingChanID: lnwire.NewShortChanIDFromInt(uint64(rand.Int63())),
|
|
OutgoingChanID: lnwire.NewShortChanIDFromInt(uint64(rand.Int63())),
|
|
AmtIn: lnwire.MilliSatoshi(rand.Int63()),
|
|
AmtOut: lnwire.MilliSatoshi(rand.Int63()),
|
|
}
|
|
}
|
|
|
|
// Now that all of our events are constructed, we'll add them to the
|
|
// database in a batched manner.
|
|
if err := log.AddForwardingEvents(events); err != nil {
|
|
t.Fatalf("unable to add events: %v", err)
|
|
}
|
|
|
|
// Because timestamps are de-duplicated when adding them in a single
|
|
// batch before they even hit the DB, we add the same events again but
|
|
// in a new batch. They now have to be de-duplicated on the DB level.
|
|
if err := log.AddForwardingEvents(events); err != nil {
|
|
t.Fatalf("unable to add second batch of events: %v", err)
|
|
}
|
|
|
|
// With all of our events added, we should be able to query for all
|
|
// events with a range of just 40 nanoseconds (2 times 20 events, all
|
|
// spaced one nanosecond apart).
|
|
eventQuery := ForwardingEventQuery{
|
|
StartTime: time.Unix(0, ts),
|
|
EndTime: time.Unix(0, ts+int64(numEvents*2)),
|
|
IndexOffset: 0,
|
|
NumMaxEvents: uint32(numEvents * 3),
|
|
}
|
|
timeSlice, err := log.Query(eventQuery)
|
|
if err != nil {
|
|
t.Fatalf("unable to query for events: %v", err)
|
|
}
|
|
|
|
// We should get exactly 40 events back.
|
|
if len(timeSlice.ForwardingEvents) != numEvents*2 {
|
|
t.Fatalf("wrong number of events: expected %v, got %v",
|
|
numEvents*2, len(timeSlice.ForwardingEvents))
|
|
}
|
|
|
|
// The timestamps should be spaced out evenly and in order.
|
|
for i := 0; i < numEvents*2; i++ {
|
|
eventTs := timeSlice.ForwardingEvents[i].Timestamp.UnixNano()
|
|
if eventTs != ts+int64(i) {
|
|
t.Fatalf("unexpected timestamp of event %d: expected "+
|
|
"%d, got %d", i, ts+int64(i), eventTs)
|
|
}
|
|
}
|
|
}
|