mirror of
https://github.com/lightningnetwork/lnd.git
synced 2024-11-19 09:53:54 +01:00
266 lines
8.1 KiB
Go
266 lines
8.1 KiB
Go
|
package channeldb
|
||
|
|
||
|
import (
|
||
|
"math/rand"
|
||
|
"reflect"
|
||
|
"testing"
|
||
|
|
||
|
"github.com/davecgh/go-spew/spew"
|
||
|
"github.com/lightningnetwork/lnd/lnwire"
|
||
|
|
||
|
"time"
|
||
|
)
|
||
|
|
||
|
// TestForwardingLogBasicStorageAndQuery tests that we're able to store and
|
||
|
// then query for items that have previously been added to the event log.
|
||
|
func TestForwardingLogBasicStorageAndQuery(t *testing.T) {
|
||
|
t.Parallel()
|
||
|
|
||
|
// First, we'll set up a test database, and use that to instantiate the
|
||
|
// forwarding event log that we'll be using for the duration of the
|
||
|
// test.
|
||
|
db, cleanUp, err := makeTestDB()
|
||
|
defer cleanUp()
|
||
|
if err != nil {
|
||
|
t.Fatalf("unable to make test db: %v", err)
|
||
|
}
|
||
|
log := ForwardingLog{
|
||
|
db: db,
|
||
|
}
|
||
|
|
||
|
initialTime := time.Unix(1234, 0)
|
||
|
timestamp := time.Unix(1234, 0)
|
||
|
|
||
|
// We'll create 100 random events, which each event being spaced 10
|
||
|
// minutes after the prior event.
|
||
|
numEvents := 100
|
||
|
events := make([]ForwardingEvent, numEvents)
|
||
|
for i := 0; i < numEvents; i++ {
|
||
|
events[i] = ForwardingEvent{
|
||
|
Timestamp: timestamp,
|
||
|
IncomingChanID: lnwire.NewShortChanIDFromInt(uint64(rand.Int63())),
|
||
|
OutgoingChanID: lnwire.NewShortChanIDFromInt(uint64(rand.Int63())),
|
||
|
AmtIn: lnwire.MilliSatoshi(rand.Int63()),
|
||
|
AmtOut: lnwire.MilliSatoshi(rand.Int63()),
|
||
|
}
|
||
|
|
||
|
timestamp = timestamp.Add(time.Minute * 10)
|
||
|
}
|
||
|
|
||
|
// Now that all of our set of events constructed, we'll add them to the
|
||
|
// database in a batch manner.
|
||
|
if err := log.AddForwardingEvents(events); err != nil {
|
||
|
t.Fatalf("unable to add events: %v", err)
|
||
|
}
|
||
|
|
||
|
// With our events added we'll now construct a basic query to retrieve
|
||
|
// all of the events.
|
||
|
eventQuery := ForwardingEventQuery{
|
||
|
StartTime: initialTime,
|
||
|
EndTime: timestamp,
|
||
|
IndexOffset: 0,
|
||
|
NumMaxEvents: 1000,
|
||
|
}
|
||
|
timeSlice, err := log.Query(eventQuery)
|
||
|
if err != nil {
|
||
|
t.Fatalf("unable to query for events: %v", err)
|
||
|
}
|
||
|
|
||
|
// The set of returned events should match identically, as they should
|
||
|
// be returned in sorted order.
|
||
|
if !reflect.DeepEqual(events, timeSlice.ForwardingEvents) {
|
||
|
t.Fatalf("event mismatch: expected %v vs %v",
|
||
|
spew.Sdump(events), spew.Sdump(timeSlice.ForwardingEvents))
|
||
|
}
|
||
|
|
||
|
// The offset index of the final entry should be numEvents, so the
|
||
|
// number of total events we've written.
|
||
|
if timeSlice.LastIndexOffset != uint32(numEvents) {
|
||
|
t.Fatalf("wrong final offset: expected %v, got %v",
|
||
|
timeSlice.LastIndexOffset, numEvents)
|
||
|
}
|
||
|
}
|
||
|
|
||
|
// TestForwardingLogQueryOptions tests that the query offset works properly. So
|
||
|
// if we add a series of events, then we should be able to seek within the
|
||
|
// timeslice accordingly. This exercises the index offset and num max event
|
||
|
// field in the query, and also the last index offset field int he response.
|
||
|
func TestForwardingLogQueryOptions(t *testing.T) {
|
||
|
t.Parallel()
|
||
|
|
||
|
// First, we'll set up a test database, and use that to instantiate the
|
||
|
// forwarding event log that we'll be using for the duration of the
|
||
|
// test.
|
||
|
db, cleanUp, err := makeTestDB()
|
||
|
defer cleanUp()
|
||
|
if err != nil {
|
||
|
t.Fatalf("unable to make test db: %v", err)
|
||
|
}
|
||
|
log := ForwardingLog{
|
||
|
db: db,
|
||
|
}
|
||
|
|
||
|
initialTime := time.Unix(1234, 0)
|
||
|
endTime := time.Unix(1234, 0)
|
||
|
|
||
|
// We'll create 20 random events, which each event being spaced 10
|
||
|
// minutes after the prior event.
|
||
|
numEvents := 20
|
||
|
events := make([]ForwardingEvent, numEvents)
|
||
|
for i := 0; i < numEvents; i++ {
|
||
|
events[i] = ForwardingEvent{
|
||
|
Timestamp: endTime,
|
||
|
IncomingChanID: lnwire.NewShortChanIDFromInt(uint64(rand.Int63())),
|
||
|
OutgoingChanID: lnwire.NewShortChanIDFromInt(uint64(rand.Int63())),
|
||
|
AmtIn: lnwire.MilliSatoshi(rand.Int63()),
|
||
|
AmtOut: lnwire.MilliSatoshi(rand.Int63()),
|
||
|
}
|
||
|
|
||
|
endTime = endTime.Add(time.Minute * 10)
|
||
|
}
|
||
|
|
||
|
// Now that all of our set of events constructed, we'll add them to the
|
||
|
// database in a batch manner.
|
||
|
if err := log.AddForwardingEvents(events); err != nil {
|
||
|
t.Fatalf("unable to add events: %v", err)
|
||
|
}
|
||
|
|
||
|
// With all of our events added, we should be able to query for the
|
||
|
// first 10 events using the max event query field.
|
||
|
eventQuery := ForwardingEventQuery{
|
||
|
StartTime: initialTime,
|
||
|
EndTime: endTime,
|
||
|
IndexOffset: 0,
|
||
|
NumMaxEvents: 10,
|
||
|
}
|
||
|
timeSlice, err := log.Query(eventQuery)
|
||
|
if err != nil {
|
||
|
t.Fatalf("unable to query for events: %v", err)
|
||
|
}
|
||
|
|
||
|
// We should get exactly 10 events back.
|
||
|
if len(timeSlice.ForwardingEvents) != 10 {
|
||
|
t.Fatalf("wrong number of events: expected %v, got %v", 10,
|
||
|
len(timeSlice.ForwardingEvents))
|
||
|
}
|
||
|
|
||
|
// The set of events returned should be the first 10 events that we
|
||
|
// added.
|
||
|
if !reflect.DeepEqual(events[:10], timeSlice.ForwardingEvents) {
|
||
|
t.Fatalf("wrong response: expected %v, got %v",
|
||
|
spew.Sdump(events[:10]),
|
||
|
spew.Sdump(timeSlice.ForwardingEvents))
|
||
|
}
|
||
|
|
||
|
// The final offset should be the exact number of events returned.
|
||
|
if timeSlice.LastIndexOffset != 10 {
|
||
|
t.Fatalf("wrong index offset: expected %v, got %v", 10,
|
||
|
timeSlice.LastIndexOffset)
|
||
|
}
|
||
|
|
||
|
// If we use the final offset to query again, then we should get 10
|
||
|
// more events, that are the last 10 events we wrote.
|
||
|
eventQuery.IndexOffset = 10
|
||
|
timeSlice, err = log.Query(eventQuery)
|
||
|
if err != nil {
|
||
|
t.Fatalf("unable to query for events: %v", err)
|
||
|
}
|
||
|
|
||
|
// We should get exactly 10 events back once again.
|
||
|
if len(timeSlice.ForwardingEvents) != 10 {
|
||
|
t.Fatalf("wrong number of events: expected %v, got %v", 10,
|
||
|
len(timeSlice.ForwardingEvents))
|
||
|
}
|
||
|
|
||
|
// The events that we got back should be the last 10 events that we
|
||
|
// wrote out.
|
||
|
if !reflect.DeepEqual(events[10:], timeSlice.ForwardingEvents) {
|
||
|
t.Fatalf("wrong response: expected %v, got %v",
|
||
|
spew.Sdump(events[10:]),
|
||
|
spew.Sdump(timeSlice.ForwardingEvents))
|
||
|
}
|
||
|
|
||
|
// Finally, the last index offset should be 20, or the number of
|
||
|
// records we've written out.
|
||
|
if timeSlice.LastIndexOffset != 20 {
|
||
|
t.Fatalf("wrong index offset: expected %v, got %v", 20,
|
||
|
timeSlice.LastIndexOffset)
|
||
|
}
|
||
|
}
|
||
|
|
||
|
// TestForwardingLogQueryLimit tests that we're able to properly limit the
|
||
|
// number of events that are returned as part of a query.
|
||
|
func TestForwardingLogQueryLimit(t *testing.T) {
|
||
|
t.Parallel()
|
||
|
|
||
|
// First, we'll set up a test database, and use that to instantiate the
|
||
|
// forwarding event log that we'll be using for the duration of the
|
||
|
// test.
|
||
|
db, cleanUp, err := makeTestDB()
|
||
|
defer cleanUp()
|
||
|
if err != nil {
|
||
|
t.Fatalf("unable to make test db: %v", err)
|
||
|
}
|
||
|
log := ForwardingLog{
|
||
|
db: db,
|
||
|
}
|
||
|
|
||
|
initialTime := time.Unix(1234, 0)
|
||
|
endTime := time.Unix(1234, 0)
|
||
|
|
||
|
// We'll create 200 random events, which each event being spaced 10
|
||
|
// minutes after the prior event.
|
||
|
numEvents := 200
|
||
|
events := make([]ForwardingEvent, numEvents)
|
||
|
for i := 0; i < numEvents; i++ {
|
||
|
events[i] = ForwardingEvent{
|
||
|
Timestamp: endTime,
|
||
|
IncomingChanID: lnwire.NewShortChanIDFromInt(uint64(rand.Int63())),
|
||
|
OutgoingChanID: lnwire.NewShortChanIDFromInt(uint64(rand.Int63())),
|
||
|
AmtIn: lnwire.MilliSatoshi(rand.Int63()),
|
||
|
AmtOut: lnwire.MilliSatoshi(rand.Int63()),
|
||
|
}
|
||
|
|
||
|
endTime = endTime.Add(time.Minute * 10)
|
||
|
}
|
||
|
|
||
|
// Now that all of our set of events constructed, we'll add them to the
|
||
|
// database in a batch manner.
|
||
|
if err := log.AddForwardingEvents(events); err != nil {
|
||
|
t.Fatalf("unable to add events: %v", err)
|
||
|
}
|
||
|
|
||
|
// Once the events have been written out, we'll issue a query over the
|
||
|
// entire range, but restrict the number of events to the first 100.
|
||
|
eventQuery := ForwardingEventQuery{
|
||
|
StartTime: initialTime,
|
||
|
EndTime: endTime,
|
||
|
IndexOffset: 0,
|
||
|
NumMaxEvents: 100,
|
||
|
}
|
||
|
timeSlice, err := log.Query(eventQuery)
|
||
|
if err != nil {
|
||
|
t.Fatalf("unable to query for events: %v", err)
|
||
|
}
|
||
|
|
||
|
// We should get exactly 100 events back.
|
||
|
if len(timeSlice.ForwardingEvents) != 100 {
|
||
|
t.Fatalf("wrong number of events: expected %v, got %v", 10,
|
||
|
len(timeSlice.ForwardingEvents))
|
||
|
}
|
||
|
|
||
|
// The set of events returned should be the first 100 events that we
|
||
|
// added.
|
||
|
if !reflect.DeepEqual(events[:100], timeSlice.ForwardingEvents) {
|
||
|
t.Fatalf("wrong response: expected %v, got %v",
|
||
|
spew.Sdump(events[:100]),
|
||
|
spew.Sdump(timeSlice.ForwardingEvents))
|
||
|
}
|
||
|
|
||
|
// The final offset should be the exact number of events returned.
|
||
|
if timeSlice.LastIndexOffset != 100 {
|
||
|
t.Fatalf("wrong index offset: expected %v, got %v", 100,
|
||
|
timeSlice.LastIndexOffset)
|
||
|
}
|
||
|
}
|