mirror of
https://github.com/lightningnetwork/lnd.git
synced 2024-11-19 18:10:34 +01:00
9a52cb6dab
In this commit, we update the router and link to support users updating the max HTLC policy for their channels. By updating these internal systems before updating the RPC server and lncli, we protect users from being shown an option that doesn't actually work.
205 lines
6.3 KiB
Go
205 lines
6.3 KiB
Go
package localchans
|
|
|
|
import (
|
|
"fmt"
|
|
"sync"
|
|
|
|
"github.com/btcsuite/btcd/wire"
|
|
"github.com/lightningnetwork/lnd/channeldb"
|
|
"github.com/lightningnetwork/lnd/discovery"
|
|
"github.com/lightningnetwork/lnd/htlcswitch"
|
|
"github.com/lightningnetwork/lnd/lnwire"
|
|
"github.com/lightningnetwork/lnd/routing"
|
|
)
|
|
|
|
// Manager manages the node's local channels. The only operation that is
|
|
// currently implemented is updating forwarding policies.
|
|
type Manager struct {
|
|
// UpdateForwardingPolicies is used by the manager to update active
|
|
// links with a new policy.
|
|
UpdateForwardingPolicies func(
|
|
chanPolicies map[wire.OutPoint]htlcswitch.ForwardingPolicy)
|
|
|
|
// PropagateChanPolicyUpdate is called to persist a new policy to disk
|
|
// and broadcast it to the network.
|
|
PropagateChanPolicyUpdate func(
|
|
edgesToUpdate []discovery.EdgeWithInfo) error
|
|
|
|
// ForAllOutgoingChannels is required to iterate over all our local
|
|
// channels.
|
|
ForAllOutgoingChannels func(cb func(*channeldb.ChannelEdgeInfo,
|
|
*channeldb.ChannelEdgePolicy) error) error
|
|
|
|
// FetchChannel is used to query local channel parameters.
|
|
FetchChannel func(chanPoint wire.OutPoint) (*channeldb.OpenChannel,
|
|
error)
|
|
|
|
// policyUpdateLock ensures that the database and the link do not fall
|
|
// out of sync if there are concurrent fee update calls. Without it,
|
|
// there is a chance that policy A updates the database, then policy B
|
|
// updates the database, then policy B updates the link, then policy A
|
|
// updates the link.
|
|
policyUpdateLock sync.Mutex
|
|
}
|
|
|
|
// UpdatePolicy updates the policy for the specified channels on disk and in the
|
|
// active links.
|
|
func (r *Manager) UpdatePolicy(newSchema routing.ChannelPolicy,
|
|
chanPoints ...wire.OutPoint) error {
|
|
|
|
r.policyUpdateLock.Lock()
|
|
defer r.policyUpdateLock.Unlock()
|
|
|
|
// First, we'll construct a set of all the channels that need to be
|
|
// updated.
|
|
chansToUpdate := make(map[wire.OutPoint]struct{})
|
|
for _, chanPoint := range chanPoints {
|
|
chansToUpdate[chanPoint] = struct{}{}
|
|
}
|
|
|
|
haveChanFilter := len(chansToUpdate) != 0
|
|
|
|
var edgesToUpdate []discovery.EdgeWithInfo
|
|
policiesToUpdate := make(map[wire.OutPoint]htlcswitch.ForwardingPolicy)
|
|
|
|
// Next, we'll loop over all the outgoing channels the router knows of.
|
|
// If we have a filter then we'll only collected those channels,
|
|
// otherwise we'll collect them all.
|
|
err := r.ForAllOutgoingChannels(func(
|
|
info *channeldb.ChannelEdgeInfo,
|
|
edge *channeldb.ChannelEdgePolicy) error {
|
|
|
|
// If we have a channel filter, and this channel isn't a part
|
|
// of it, then we'll skip it.
|
|
_, ok := chansToUpdate[info.ChannelPoint]
|
|
if !ok && haveChanFilter {
|
|
return nil
|
|
}
|
|
|
|
// Apply the new policy to the edge.
|
|
err := r.updateEdge(info.ChannelPoint, edge, newSchema)
|
|
if err != nil {
|
|
return nil
|
|
}
|
|
|
|
// Add updated edge to list of edges to send to gossiper.
|
|
edgesToUpdate = append(edgesToUpdate, discovery.EdgeWithInfo{
|
|
Info: info,
|
|
Edge: edge,
|
|
})
|
|
|
|
// Add updated policy to list of policies to send to switch.
|
|
policiesToUpdate[info.ChannelPoint] = htlcswitch.ForwardingPolicy{
|
|
BaseFee: edge.FeeBaseMSat,
|
|
FeeRate: edge.FeeProportionalMillionths,
|
|
TimeLockDelta: uint32(edge.TimeLockDelta),
|
|
MinHTLC: edge.MinHTLC,
|
|
MaxHTLC: edge.MaxHTLC,
|
|
}
|
|
|
|
return nil
|
|
})
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
// Commit the policy updates to disk and broadcast to the network. We
|
|
// validated the new policy above, so we expect no validation errors. If
|
|
// this would happen because of a bug, the link policy will be
|
|
// desynchronized. It is currently not possible to atomically commit
|
|
// multiple edge updates.
|
|
err = r.PropagateChanPolicyUpdate(edgesToUpdate)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
// Update active links.
|
|
r.UpdateForwardingPolicies(policiesToUpdate)
|
|
|
|
return nil
|
|
}
|
|
|
|
// updateEdge updates the given edge with the new schema.
|
|
func (r *Manager) updateEdge(chanPoint wire.OutPoint,
|
|
edge *channeldb.ChannelEdgePolicy,
|
|
newSchema routing.ChannelPolicy) error {
|
|
|
|
// Update forwarding fee scheme and required time lock delta.
|
|
edge.FeeBaseMSat = newSchema.BaseFee
|
|
edge.FeeProportionalMillionths = lnwire.MilliSatoshi(
|
|
newSchema.FeeRate,
|
|
)
|
|
edge.TimeLockDelta = uint16(newSchema.TimeLockDelta)
|
|
|
|
// Retrieve negotiated channel htlc amt limits.
|
|
amtMin, amtMax, err := r.getHtlcAmtLimits(chanPoint)
|
|
if err != nil {
|
|
return nil
|
|
}
|
|
|
|
// We now update the edge max htlc value.
|
|
switch {
|
|
|
|
// If a non-zero max htlc was specified, use it to update the edge.
|
|
// Otherwise keep the value unchanged.
|
|
case newSchema.MaxHTLC != 0:
|
|
edge.MaxHTLC = newSchema.MaxHTLC
|
|
|
|
// If this edge still doesn't have a max htlc set, set it to the max.
|
|
// This is an on-the-fly migration.
|
|
case !edge.MessageFlags.HasMaxHtlc():
|
|
edge.MaxHTLC = amtMax
|
|
|
|
// If this edge has a max htlc that exceeds what the channel can
|
|
// actually carry, correct it now. This can happen, because we
|
|
// previously set the max htlc to the channel capacity.
|
|
case edge.MaxHTLC > amtMax:
|
|
edge.MaxHTLC = amtMax
|
|
}
|
|
|
|
// If the MaxHtlc flag wasn't already set, we can set it now.
|
|
edge.MessageFlags |= lnwire.ChanUpdateOptionMaxHtlc
|
|
|
|
// Validate htlc amount constraints.
|
|
switch {
|
|
case edge.MinHTLC < amtMin:
|
|
return fmt.Errorf("min htlc amount of %v msat is below "+
|
|
"min htlc parameter of %v msat for channel %v",
|
|
edge.MinHTLC, amtMin,
|
|
chanPoint)
|
|
|
|
case edge.MaxHTLC > amtMax:
|
|
return fmt.Errorf("max htlc size of %v msat is above "+
|
|
"max pending amount of %v msat for channel %v",
|
|
edge.MaxHTLC, amtMax, chanPoint)
|
|
|
|
case edge.MinHTLC > edge.MaxHTLC:
|
|
return fmt.Errorf("min_htlc %v greater than max_htlc %v",
|
|
edge.MinHTLC, edge.MaxHTLC)
|
|
}
|
|
|
|
// Clear signature to help prevent usage of the previous signature.
|
|
edge.SetSigBytes(nil)
|
|
|
|
return nil
|
|
}
|
|
|
|
// getHtlcAmtLimits retrieves the negotiated channel min and max htlc amount
|
|
// constraints.
|
|
func (r *Manager) getHtlcAmtLimits(chanPoint wire.OutPoint) (
|
|
lnwire.MilliSatoshi, lnwire.MilliSatoshi, error) {
|
|
|
|
ch, err := r.FetchChannel(chanPoint)
|
|
if err != nil {
|
|
return 0, 0, err
|
|
}
|
|
|
|
// The max htlc policy field must be less than or equal to the channel
|
|
// capacity AND less than or equal to the max in-flight HTLC value.
|
|
// Since the latter is always less than or equal to the former, just
|
|
// return the max in-flight value.
|
|
maxAmt := ch.LocalChanCfg.ChannelConstraints.MaxPendingAmount
|
|
|
|
return ch.LocalChanCfg.MinHTLC, maxAmt, nil
|
|
}
|