Merge pull request #2769 from TheBlueMatt/2023-12-2314-cleanups-3

Post-#2314 Cleanups #3
This commit is contained in:
Wilmer Paulino 2023-12-04 10:09:52 -08:00 committed by GitHub
commit 4d66cce487
No known key found for this signature in database
GPG key ID: 4AEE18F83AFDEB23
3 changed files with 55 additions and 18 deletions

View file

@ -599,6 +599,8 @@ pub(crate) struct ShutdownResult {
/// An unbroadcasted batch funding transaction id. The closure of this channel should be /// An unbroadcasted batch funding transaction id. The closure of this channel should be
/// propagated to the remainder of the batch. /// propagated to the remainder of the batch.
pub(crate) unbroadcasted_batch_funding_txid: Option<Txid>, pub(crate) unbroadcasted_batch_funding_txid: Option<Txid>,
pub(crate) channel_id: ChannelId,
pub(crate) counterparty_node_id: PublicKey,
} }
/// If the majority of the channels funds are to the fundee and the initiator holds only just /// If the majority of the channels funds are to the fundee and the initiator holds only just
@ -2164,6 +2166,8 @@ impl<SP: Deref> ChannelContext<SP> where SP::Target: SignerProvider {
monitor_update, monitor_update,
dropped_outbound_htlcs, dropped_outbound_htlcs,
unbroadcasted_batch_funding_txid, unbroadcasted_batch_funding_txid,
channel_id: self.channel_id,
counterparty_node_id: self.counterparty_node_id,
} }
} }
@ -4707,6 +4711,8 @@ impl<SP: Deref> Channel<SP> where
monitor_update: None, monitor_update: None,
dropped_outbound_htlcs: Vec::new(), dropped_outbound_htlcs: Vec::new(),
unbroadcasted_batch_funding_txid: self.context.unbroadcasted_batch_funding_txid(), unbroadcasted_batch_funding_txid: self.context.unbroadcasted_batch_funding_txid(),
channel_id: self.context.channel_id,
counterparty_node_id: self.context.counterparty_node_id,
}; };
let tx = self.build_signed_closing_transaction(&mut closing_tx, &msg.signature, &sig); let tx = self.build_signed_closing_transaction(&mut closing_tx, &msg.signature, &sig);
self.context.channel_state = ChannelState::ShutdownComplete as u32; self.context.channel_state = ChannelState::ShutdownComplete as u32;
@ -4735,6 +4741,8 @@ impl<SP: Deref> Channel<SP> where
monitor_update: None, monitor_update: None,
dropped_outbound_htlcs: Vec::new(), dropped_outbound_htlcs: Vec::new(),
unbroadcasted_batch_funding_txid: self.context.unbroadcasted_batch_funding_txid(), unbroadcasted_batch_funding_txid: self.context.unbroadcasted_batch_funding_txid(),
channel_id: self.context.channel_id,
counterparty_node_id: self.context.counterparty_node_id,
}; };
self.context.channel_state = ChannelState::ShutdownComplete as u32; self.context.channel_state = ChannelState::ShutdownComplete as u32;
self.context.update_time_counter += 1; self.context.update_time_counter += 1;
@ -5913,6 +5921,8 @@ impl<SP: Deref> Channel<SP> where
monitor_update: None, monitor_update: None,
dropped_outbound_htlcs: Vec::new(), dropped_outbound_htlcs: Vec::new(),
unbroadcasted_batch_funding_txid: self.context.unbroadcasted_batch_funding_txid(), unbroadcasted_batch_funding_txid: self.context.unbroadcasted_batch_funding_txid(),
channel_id: self.context.channel_id,
counterparty_node_id: self.context.counterparty_node_id,
}; };
self.context.channel_state = ChannelState::ShutdownComplete as u32; self.context.channel_state = ChannelState::ShutdownComplete as u32;
Some(shutdown_result) Some(shutdown_result)

View file

@ -2808,7 +2808,10 @@ where
debug_assert_ne!(peer.held_by_thread(), LockHeldState::HeldByThread); debug_assert_ne!(peer.held_by_thread(), LockHeldState::HeldByThread);
} }
log_debug!(self.logger, "Finishing closure of channel with {} HTLCs to fail", shutdown_res.dropped_outbound_htlcs.len()); let logger = WithContext::from(
&self.logger, Some(shutdown_res.counterparty_node_id), Some(shutdown_res.channel_id),
);
log_debug!(logger, "Finishing closure of channel with {} HTLCs to fail", shutdown_res.dropped_outbound_htlcs.len());
for htlc_source in shutdown_res.dropped_outbound_htlcs.drain(..) { for htlc_source in shutdown_res.dropped_outbound_htlcs.drain(..) {
let (source, payment_hash, counterparty_node_id, channel_id) = htlc_source; let (source, payment_hash, counterparty_node_id, channel_id) = htlc_source;
let reason = HTLCFailReason::from_failure_code(0x4000 | 8); let reason = HTLCFailReason::from_failure_code(0x4000 | 8);
@ -2967,7 +2970,7 @@ where
} }
fn decode_update_add_htlc_onion( fn decode_update_add_htlc_onion(
&self, msg: &msgs::UpdateAddHTLC &self, msg: &msgs::UpdateAddHTLC, counterparty_node_id: &PublicKey,
) -> Result< ) -> Result<
(onion_utils::Hop, [u8; 32], Option<Result<PublicKey, secp256k1::Error>>), HTLCFailureMsg (onion_utils::Hop, [u8; 32], Option<Result<PublicKey, secp256k1::Error>>), HTLCFailureMsg
> { > {
@ -2986,7 +2989,7 @@ where
($msg: expr, $err_code: expr, $data: expr) => { ($msg: expr, $err_code: expr, $data: expr) => {
{ {
log_info!( log_info!(
WithContext::from(&self.logger, None, Some(msg.channel_id)), WithContext::from(&self.logger, Some(*counterparty_node_id), Some(msg.channel_id)),
"Failed to accept/forward incoming HTLC: {}", $msg "Failed to accept/forward incoming HTLC: {}", $msg
); );
let (err_code, err_data) = if is_blinded { let (err_code, err_data) = if is_blinded {
@ -3133,13 +3136,15 @@ where
} }
fn construct_pending_htlc_status<'a>( fn construct_pending_htlc_status<'a>(
&self, msg: &msgs::UpdateAddHTLC, shared_secret: [u8; 32], decoded_hop: onion_utils::Hop, &self, msg: &msgs::UpdateAddHTLC, counterparty_node_id: &PublicKey, shared_secret: [u8; 32],
allow_underpay: bool, next_packet_pubkey_opt: Option<Result<PublicKey, secp256k1::Error>> decoded_hop: onion_utils::Hop, allow_underpay: bool,
next_packet_pubkey_opt: Option<Result<PublicKey, secp256k1::Error>>,
) -> PendingHTLCStatus { ) -> PendingHTLCStatus {
macro_rules! return_err { macro_rules! return_err {
($msg: expr, $err_code: expr, $data: expr) => { ($msg: expr, $err_code: expr, $data: expr) => {
{ {
log_info!(WithContext::from(&self.logger, None, Some(msg.channel_id)), "Failed to accept/forward incoming HTLC: {}", $msg); let logger = WithContext::from(&self.logger, Some(*counterparty_node_id), Some(msg.channel_id));
log_info!(logger, "Failed to accept/forward incoming HTLC: {}", $msg);
return PendingHTLCStatus::Fail(HTLCFailureMsg::Relay(msgs::UpdateFailHTLC { return PendingHTLCStatus::Fail(HTLCFailureMsg::Relay(msgs::UpdateFailHTLC {
channel_id: msg.channel_id, channel_id: msg.channel_id,
htlc_id: msg.htlc_id, htlc_id: msg.htlc_id,
@ -3276,23 +3281,33 @@ where
} = args; } = args;
// The top-level caller should hold the total_consistency_lock read lock. // The top-level caller should hold the total_consistency_lock read lock.
debug_assert!(self.total_consistency_lock.try_write().is_err()); debug_assert!(self.total_consistency_lock.try_write().is_err());
log_trace!(WithContext::from(&self.logger, Some(path.hops.first().unwrap().pubkey), None),
"Attempting to send payment with payment hash {} along path with next hop {}",
payment_hash, path.hops.first().unwrap().short_channel_id);
let prng_seed = self.entropy_source.get_secure_random_bytes(); let prng_seed = self.entropy_source.get_secure_random_bytes();
let session_priv = SecretKey::from_slice(&session_priv_bytes[..]).expect("RNG is busted"); let session_priv = SecretKey::from_slice(&session_priv_bytes[..]).expect("RNG is busted");
let (onion_packet, htlc_msat, htlc_cltv) = onion_utils::create_payment_onion( let (onion_packet, htlc_msat, htlc_cltv) = onion_utils::create_payment_onion(
&self.secp_ctx, &path, &session_priv, total_value, recipient_onion, cur_height, &self.secp_ctx, &path, &session_priv, total_value, recipient_onion, cur_height,
payment_hash, keysend_preimage, prng_seed payment_hash, keysend_preimage, prng_seed
)?; ).map_err(|e| {
let logger = WithContext::from(&self.logger, Some(path.hops.first().unwrap().pubkey), None);
log_error!(logger, "Failed to build an onion for path for payment hash {}", payment_hash);
e
})?;
let err: Result<(), _> = loop { let err: Result<(), _> = loop {
let (counterparty_node_id, id) = match self.short_to_chan_info.read().unwrap().get(&path.hops.first().unwrap().short_channel_id) { let (counterparty_node_id, id) = match self.short_to_chan_info.read().unwrap().get(&path.hops.first().unwrap().short_channel_id) {
None => return Err(APIError::ChannelUnavailable{err: "No channel available with first hop!".to_owned()}), None => {
let logger = WithContext::from(&self.logger, Some(path.hops.first().unwrap().pubkey), None);
log_error!(logger, "Failed to find first-hop for payment hash {}", payment_hash);
return Err(APIError::ChannelUnavailable{err: "No channel available with first hop!".to_owned()})
},
Some((cp_id, chan_id)) => (cp_id.clone(), chan_id.clone()), Some((cp_id, chan_id)) => (cp_id.clone(), chan_id.clone()),
}; };
let logger = WithContext::from(&self.logger, Some(counterparty_node_id), Some(id));
log_trace!(logger,
"Attempting to send payment with payment hash {} along path with next hop {}",
payment_hash, path.hops.first().unwrap().short_channel_id);
let per_peer_state = self.per_peer_state.read().unwrap(); let per_peer_state = self.per_peer_state.read().unwrap();
let peer_state_mutex = per_peer_state.get(&counterparty_node_id) let peer_state_mutex = per_peer_state.get(&counterparty_node_id)
.ok_or_else(|| APIError::ChannelUnavailable{err: "No peer matching the path's first hop found!".to_owned() })?; .ok_or_else(|| APIError::ChannelUnavailable{err: "No peer matching the path's first hop found!".to_owned() })?;
@ -4044,7 +4059,8 @@ where
None => { None => {
let error = format!("Channel with id {} not found for the passed counterparty node_id {}", let error = format!("Channel with id {} not found for the passed counterparty node_id {}",
next_hop_channel_id, next_node_id); next_hop_channel_id, next_node_id);
log_error!(self.logger, "{} when attempting to forward intercepted HTLC", error); let logger = WithContext::from(&self.logger, Some(next_node_id), Some(*next_hop_channel_id));
log_error!(logger, "{} when attempting to forward intercepted HTLC", error);
return Err(APIError::ChannelUnavailable { return Err(APIError::ChannelUnavailable {
err: error err: error
}) })
@ -4132,6 +4148,7 @@ where
for (short_chan_id, mut pending_forwards) in forward_htlcs { for (short_chan_id, mut pending_forwards) in forward_htlcs {
if short_chan_id != 0 { if short_chan_id != 0 {
let mut forwarding_counterparty = None;
macro_rules! forwarding_channel_not_found { macro_rules! forwarding_channel_not_found {
() => { () => {
for forward_info in pending_forwards.drain(..) { for forward_info in pending_forwards.drain(..) {
@ -4145,7 +4162,8 @@ where
}) => { }) => {
macro_rules! failure_handler { macro_rules! failure_handler {
($msg: expr, $err_code: expr, $err_data: expr, $phantom_ss: expr, $next_hop_unknown: expr) => { ($msg: expr, $err_code: expr, $err_data: expr, $phantom_ss: expr, $next_hop_unknown: expr) => {
log_info!(self.logger, "Failed to accept/forward incoming HTLC: {}", $msg); let logger = WithContext::from(&self.logger, forwarding_counterparty, Some(prev_funding_outpoint.to_channel_id()));
log_info!(logger, "Failed to accept/forward incoming HTLC: {}", $msg);
let htlc_source = HTLCSource::PreviousHopData(HTLCPreviousHopData { let htlc_source = HTLCSource::PreviousHopData(HTLCPreviousHopData {
short_channel_id: prev_short_channel_id, short_channel_id: prev_short_channel_id,
@ -4244,6 +4262,7 @@ where
continue; continue;
} }
}; };
forwarding_counterparty = Some(counterparty_node_id);
let per_peer_state = self.per_peer_state.read().unwrap(); let per_peer_state = self.per_peer_state.read().unwrap();
let peer_state_mutex_opt = per_peer_state.get(&counterparty_node_id); let peer_state_mutex_opt = per_peer_state.get(&counterparty_node_id);
if peer_state_mutex_opt.is_none() { if peer_state_mutex_opt.is_none() {
@ -5347,6 +5366,7 @@ where
} }
if valid_mpp { if valid_mpp {
for htlc in sources.drain(..) { for htlc in sources.drain(..) {
let prev_hop_chan_id = htlc.prev_hop.outpoint.to_channel_id();
if let Err((pk, err)) = self.claim_funds_from_hop( if let Err((pk, err)) = self.claim_funds_from_hop(
htlc.prev_hop, payment_preimage, htlc.prev_hop, payment_preimage,
|_, definitely_duplicate| { |_, definitely_duplicate| {
@ -5357,6 +5377,7 @@ where
if let msgs::ErrorAction::IgnoreError = err.err.action { if let msgs::ErrorAction::IgnoreError = err.err.action {
// We got a temporary failure updating monitor, but will claim the // We got a temporary failure updating monitor, but will claim the
// HTLC when the monitor updating is restored (or on chain). // HTLC when the monitor updating is restored (or on chain).
let logger = WithContext::from(&self.logger, None, Some(prev_hop_chan_id));
log_error!(self.logger, "Temporary failure claiming HTLC, treating as success: {}", err.err.err); log_error!(self.logger, "Temporary failure claiming HTLC, treating as success: {}", err.err.err);
} else { errs.push((pk, err)); } } else { errs.push((pk, err)); }
} }
@ -6455,7 +6476,7 @@ where
// Note that the ChannelManager is NOT re-persisted on disk after this (unless we error // Note that the ChannelManager is NOT re-persisted on disk after this (unless we error
// closing a channel), so any changes are likely to be lost on restart! // closing a channel), so any changes are likely to be lost on restart!
let decoded_hop_res = self.decode_update_add_htlc_onion(msg); let decoded_hop_res = self.decode_update_add_htlc_onion(msg, counterparty_node_id);
let per_peer_state = self.per_peer_state.read().unwrap(); let per_peer_state = self.per_peer_state.read().unwrap();
let peer_state_mutex = per_peer_state.get(counterparty_node_id) let peer_state_mutex = per_peer_state.get(counterparty_node_id)
.ok_or_else(|| { .ok_or_else(|| {
@ -6469,8 +6490,10 @@ where
if let ChannelPhase::Funded(chan) = chan_phase_entry.get_mut() { if let ChannelPhase::Funded(chan) = chan_phase_entry.get_mut() {
let pending_forward_info = match decoded_hop_res { let pending_forward_info = match decoded_hop_res {
Ok((next_hop, shared_secret, next_packet_pk_opt)) => Ok((next_hop, shared_secret, next_packet_pk_opt)) =>
self.construct_pending_htlc_status(msg, shared_secret, next_hop, self.construct_pending_htlc_status(
chan.context.config().accept_underpaying_htlcs, next_packet_pk_opt), msg, counterparty_node_id, shared_secret, next_hop,
chan.context.config().accept_underpaying_htlcs, next_packet_pk_opt,
),
Err(e) => PendingHTLCStatus::Fail(e) Err(e) => PendingHTLCStatus::Fail(e)
}; };
let create_pending_htlc_status = |chan: &Channel<SP>, pending_forward_info: PendingHTLCStatus, error_code: u16| { let create_pending_htlc_status = |chan: &Channel<SP>, pending_forward_info: PendingHTLCStatus, error_code: u16| {
@ -10014,7 +10037,7 @@ where
log_error!(logger, "A ChannelManager is stale compared to the current ChannelMonitor!"); log_error!(logger, "A ChannelManager is stale compared to the current ChannelMonitor!");
log_error!(logger, " The channel will be force-closed and the latest commitment transaction from the ChannelMonitor broadcast."); log_error!(logger, " The channel will be force-closed and the latest commitment transaction from the ChannelMonitor broadcast.");
if channel.context.get_latest_monitor_update_id() < monitor.get_latest_update_id() { if channel.context.get_latest_monitor_update_id() < monitor.get_latest_update_id() {
log_error!(args.logger, " The ChannelMonitor for channel {} is at update_id {} but the ChannelManager is at update_id {}.", log_error!(logger, " The ChannelMonitor for channel {} is at update_id {} but the ChannelManager is at update_id {}.",
&channel.context.channel_id(), monitor.get_latest_update_id(), channel.context.get_latest_monitor_update_id()); &channel.context.channel_id(), monitor.get_latest_update_id(), channel.context.get_latest_monitor_update_id());
} }
if channel.get_cur_holder_commitment_transaction_number() > monitor.get_cur_holder_commitment_number() { if channel.get_cur_holder_commitment_transaction_number() > monitor.get_cur_holder_commitment_number() {
@ -10410,7 +10433,7 @@ where
let counterparty_opt = id_to_peer.get(&monitor.get_funding_txo().0.to_channel_id()); let counterparty_opt = id_to_peer.get(&monitor.get_funding_txo().0.to_channel_id());
let chan_id = monitor.get_funding_txo().0.to_channel_id(); let chan_id = monitor.get_funding_txo().0.to_channel_id();
if counterparty_opt.is_none() { if counterparty_opt.is_none() {
let logger = WithContext::from(&args.logger, None, Some(chan_id)); let logger = WithChannelMonitor::from(&args.logger, monitor);
for (htlc_source, (htlc, _)) in monitor.get_pending_or_resolved_outbound_htlcs() { for (htlc_source, (htlc, _)) in monitor.get_pending_or_resolved_outbound_htlcs() {
if let HTLCSource::OutboundRoute { payment_id, session_priv, path, .. } = htlc_source { if let HTLCSource::OutboundRoute { payment_id, session_priv, path, .. } = htlc_source {
if path.hops.is_empty() { if path.hops.is_empty() {

View file

@ -98,6 +98,10 @@ pub struct Record<'a> {
/// The verbosity level of the message. /// The verbosity level of the message.
pub level: Level, pub level: Level,
/// The node id of the peer pertaining to the logged record. /// The node id of the peer pertaining to the logged record.
///
/// Note that in some cases a [`Self::channel_id`] may be filled in but this may still be
/// `None`, depending on if the peer information is readily available in LDK when the log is
/// generated.
pub peer_id: Option<PublicKey>, pub peer_id: Option<PublicKey>,
/// The channel id of the channel pertaining to the logged record. May be a temporary id before /// The channel id of the channel pertaining to the logged record. May be a temporary id before
/// the channel has been funded. /// the channel has been funded.