mirror of
https://github.com/ElementsProject/lightning.git
synced 2024-11-19 09:54:16 +01:00
bcf3cef96c
Make sure it sees disconnect before reconnect, otherwise the next command fails since we're now disconnected. Signed-off-by: Rusty Russell <rusty@rustcorp.com.au>
2866 lines
86 KiB
C
2866 lines
86 KiB
C
#include "config.h"
|
|
#include <arpa/inet.h>
|
|
#include <bitcoin/feerate.h>
|
|
#include <bitcoin/script.h>
|
|
#include <bitcoin/tx.h>
|
|
#include <ccan/array_size/array_size.h>
|
|
#include <ccan/cast/cast.h>
|
|
#include <ccan/io/io.h>
|
|
#include <ccan/mem/mem.h>
|
|
#include <ccan/noerr/noerr.h>
|
|
#include <ccan/str/str.h>
|
|
#include <ccan/take/take.h>
|
|
#include <ccan/tal/str/str.h>
|
|
#include <channeld/channeld_wiregen.h>
|
|
#include <common/addr.h>
|
|
#include <common/closing_fee.h>
|
|
#include <common/configdir.h>
|
|
#include <common/dev_disconnect.h>
|
|
#include <common/features.h>
|
|
#include <common/htlc_trim.h>
|
|
#include <common/initial_commit_tx.h>
|
|
#include <common/json_command.h>
|
|
#include <common/json_helpers.h>
|
|
#include <common/json_tok.h>
|
|
#include <common/jsonrpc_errors.h>
|
|
#include <common/key_derive.h>
|
|
#include <common/param.h>
|
|
#include <common/shutdown_scriptpubkey.h>
|
|
#include <common/status.h>
|
|
#include <common/timeout.h>
|
|
#include <common/type_to_string.h>
|
|
#include <common/utils.h>
|
|
#include <common/version.h>
|
|
#include <common/wire_error.h>
|
|
#include <connectd/connectd_wiregen.h>
|
|
#include <errno.h>
|
|
#include <fcntl.h>
|
|
#include <gossipd/gossipd_wiregen.h>
|
|
#include <hsmd/hsmd_wiregen.h>
|
|
#include <inttypes.h>
|
|
#include <lightningd/bitcoind.h>
|
|
#include <lightningd/chaintopology.h>
|
|
#include <lightningd/channel.h>
|
|
#include <lightningd/channel_control.h>
|
|
#include <lightningd/closing_control.h>
|
|
#include <lightningd/connect_control.h>
|
|
#include <lightningd/dual_open_control.h>
|
|
#include <lightningd/hsm_control.h>
|
|
#include <lightningd/json.h>
|
|
#include <lightningd/jsonrpc.h>
|
|
#include <lightningd/lightningd.h>
|
|
#include <lightningd/log.h>
|
|
#include <lightningd/memdump.h>
|
|
#include <lightningd/notification.h>
|
|
#include <lightningd/onchain_control.h>
|
|
#include <lightningd/opening_common.h>
|
|
#include <lightningd/opening_control.h>
|
|
#include <lightningd/options.h>
|
|
#include <lightningd/peer_control.h>
|
|
#include <lightningd/peer_fd.h>
|
|
#include <lightningd/peer_htlcs.h>
|
|
#include <lightningd/plugin_hook.h>
|
|
#include <lightningd/subd.h>
|
|
#include <limits.h>
|
|
#include <onchaind/onchaind_wiregen.h>
|
|
#include <openingd/dualopend_wiregen.h>
|
|
#include <openingd/openingd_wiregen.h>
|
|
#include <stdlib.h>
|
|
#include <unistd.h>
|
|
#include <wally_bip32.h>
|
|
#include <wire/onion_wire.h>
|
|
#include <wire/wire_sync.h>
|
|
|
|
static void destroy_peer(struct peer *peer)
|
|
{
|
|
list_del_from(&peer->ld->peers, &peer->list);
|
|
}
|
|
|
|
static void peer_update_features(struct peer *peer,
|
|
const u8 *their_features TAKES)
|
|
{
|
|
tal_free(peer->their_features);
|
|
peer->their_features = tal_dup_talarr(peer, u8, their_features);
|
|
}
|
|
|
|
struct peer *new_peer(struct lightningd *ld, u64 dbid,
|
|
const struct node_id *id,
|
|
const struct wireaddr_internal *addr,
|
|
bool connected_incoming)
|
|
{
|
|
/* We are owned by our channels, and freed manually by destroy_channel */
|
|
struct peer *peer = tal(NULL, struct peer);
|
|
|
|
peer->ld = ld;
|
|
peer->dbid = dbid;
|
|
peer->id = *id;
|
|
peer->uncommitted_channel = NULL;
|
|
peer->addr = *addr;
|
|
peer->connected_incoming = connected_incoming;
|
|
peer->their_features = NULL;
|
|
list_head_init(&peer->channels);
|
|
peer->direction = node_id_idx(&peer->ld->id, &peer->id);
|
|
peer->is_connected = false;
|
|
#if DEVELOPER
|
|
peer->ignore_htlcs = false;
|
|
#endif
|
|
|
|
list_add_tail(&ld->peers, &peer->list);
|
|
tal_add_destructor(peer, destroy_peer);
|
|
return peer;
|
|
}
|
|
|
|
static void delete_peer(struct peer *peer)
|
|
{
|
|
assert(list_empty(&peer->channels));
|
|
assert(!peer->uncommitted_channel);
|
|
/* If it only ever existed because of uncommitted channel, it won't
|
|
* be in the database */
|
|
if (peer->dbid != 0)
|
|
wallet_peer_delete(peer->ld->wallet, peer->dbid);
|
|
tal_free(peer);
|
|
}
|
|
|
|
/* Last one out deletes peer. */
|
|
void maybe_delete_peer(struct peer *peer)
|
|
{
|
|
if (!list_empty(&peer->channels))
|
|
return;
|
|
if (peer->uncommitted_channel) {
|
|
/* This isn't sufficient to keep it in db! */
|
|
if (peer->dbid != 0) {
|
|
wallet_peer_delete(peer->ld->wallet, peer->dbid);
|
|
peer->dbid = 0;
|
|
}
|
|
return;
|
|
}
|
|
delete_peer(peer);
|
|
}
|
|
|
|
struct peer *find_peer_by_dbid(struct lightningd *ld, u64 dbid)
|
|
{
|
|
struct peer *p;
|
|
|
|
list_for_each(&ld->peers, p, list)
|
|
if (p->dbid == dbid)
|
|
return p;
|
|
return NULL;
|
|
}
|
|
|
|
struct peer *peer_by_id(struct lightningd *ld, const struct node_id *id)
|
|
{
|
|
struct peer *p;
|
|
|
|
list_for_each(&ld->peers, p, list)
|
|
if (node_id_eq(&p->id, id))
|
|
return p;
|
|
return NULL;
|
|
}
|
|
|
|
struct peer *peer_from_json(struct lightningd *ld,
|
|
const char *buffer,
|
|
const jsmntok_t *peeridtok)
|
|
{
|
|
struct node_id peerid;
|
|
|
|
if (!json_to_node_id(buffer, peeridtok, &peerid))
|
|
return NULL;
|
|
|
|
return peer_by_id(ld, &peerid);
|
|
}
|
|
|
|
u8 *p2wpkh_for_keyidx(const tal_t *ctx, struct lightningd *ld, u64 keyidx)
|
|
{
|
|
struct pubkey shutdownkey;
|
|
|
|
if (!bip32_pubkey(ld->wallet->bip32_base, &shutdownkey, keyidx))
|
|
return NULL;
|
|
|
|
return scriptpubkey_p2wpkh(ctx, &shutdownkey);
|
|
}
|
|
|
|
static void sign_last_tx(struct channel *channel,
|
|
struct bitcoin_tx *last_tx,
|
|
struct bitcoin_signature *last_sig)
|
|
{
|
|
struct lightningd *ld = channel->peer->ld;
|
|
struct bitcoin_signature sig;
|
|
u8 *msg, **witness;
|
|
|
|
u64 commit_index = channel->next_index[LOCAL] - 1;
|
|
|
|
assert(!last_tx->wtx->inputs[0].witness);
|
|
msg = towire_hsmd_sign_commitment_tx(tmpctx,
|
|
&channel->peer->id,
|
|
channel->dbid,
|
|
last_tx,
|
|
&channel->channel_info
|
|
.remote_fundingkey,
|
|
commit_index);
|
|
|
|
if (!wire_sync_write(ld->hsm_fd, take(msg)))
|
|
fatal("Could not write to HSM: %s", strerror(errno));
|
|
|
|
msg = wire_sync_read(tmpctx, ld->hsm_fd);
|
|
if (!fromwire_hsmd_sign_commitment_tx_reply(msg, &sig))
|
|
fatal("HSM gave bad sign_commitment_tx_reply %s",
|
|
tal_hex(tmpctx, msg));
|
|
|
|
witness =
|
|
bitcoin_witness_2of2(last_tx, last_sig,
|
|
&sig, &channel->channel_info.remote_fundingkey,
|
|
&channel->local_funding_pubkey);
|
|
|
|
bitcoin_tx_input_set_witness(last_tx, 0, take(witness));
|
|
}
|
|
|
|
static void remove_sig(struct bitcoin_tx *signed_tx)
|
|
{
|
|
bitcoin_tx_input_set_witness(signed_tx, 0, NULL);
|
|
}
|
|
|
|
static bool invalid_last_tx(const struct bitcoin_tx *tx)
|
|
{
|
|
/* This problem goes back further, but was discovered just before the
|
|
* 0.7.1 release. */
|
|
#ifdef COMPAT_V070
|
|
/* Old bug had commitment txs with no outputs; bitcoin_txid asserts. */
|
|
return tx->wtx->num_outputs == 0;
|
|
#else
|
|
return false;
|
|
#endif
|
|
}
|
|
|
|
static void sign_and_send_last(struct lightningd *ld,
|
|
struct channel *channel,
|
|
struct bitcoin_tx *last_tx,
|
|
struct bitcoin_signature *last_sig)
|
|
{
|
|
struct bitcoin_txid txid;
|
|
|
|
sign_last_tx(channel, last_tx, last_sig);
|
|
bitcoin_txid(last_tx, &txid);
|
|
wallet_transaction_add(ld->wallet, last_tx->wtx, 0, 0);
|
|
wallet_transaction_annotate(ld->wallet, &txid,
|
|
channel->last_tx_type,
|
|
channel->dbid);
|
|
|
|
/* Keep broadcasting until we say stop (can fail due to dup,
|
|
* if they beat us to the broadcast). */
|
|
broadcast_tx(ld->topology, channel, last_tx, NULL);
|
|
|
|
remove_sig(last_tx);
|
|
}
|
|
|
|
void drop_to_chain(struct lightningd *ld, struct channel *channel,
|
|
bool cooperative)
|
|
{
|
|
struct channel_inflight *inflight;
|
|
/* BOLT #2:
|
|
*
|
|
* - if `next_revocation_number` is greater than expected
|
|
* above, AND `your_last_per_commitment_secret` is correct for that
|
|
* `next_revocation_number` minus 1:
|
|
* - MUST NOT broadcast its commitment transaction.
|
|
*/
|
|
if (channel->future_per_commitment_point && !cooperative) {
|
|
log_broken(channel->log,
|
|
"Cannot broadcast our commitment tx:"
|
|
" they have a future one");
|
|
} else if (invalid_last_tx(channel->last_tx)) {
|
|
log_broken(channel->log,
|
|
"Cannot broadcast our commitment tx:"
|
|
" it's invalid! (ancient channel?)");
|
|
} else {
|
|
/* We need to drop *every* commitment transaction to chain */
|
|
if (!cooperative && !list_empty(&channel->inflights)) {
|
|
list_for_each(&channel->inflights, inflight, list)
|
|
sign_and_send_last(ld, channel,
|
|
inflight->last_tx,
|
|
&inflight->last_sig);
|
|
} else
|
|
sign_and_send_last(ld, channel, channel->last_tx,
|
|
&channel->last_sig);
|
|
}
|
|
|
|
resolve_close_command(ld, channel, cooperative);
|
|
}
|
|
|
|
void channel_errmsg(struct channel *channel,
|
|
struct peer_fd *peer_fd,
|
|
const struct channel_id *channel_id UNUSED,
|
|
const char *desc,
|
|
bool warning,
|
|
const u8 *err_for_them)
|
|
{
|
|
/* Clean up any in-progress open attempts */
|
|
channel_cleanup_commands(channel, desc);
|
|
|
|
if (channel_unsaved(channel)) {
|
|
log_info(channel->log, "%s", "Unsaved peer failed."
|
|
" Disconnecting and deleting channel.");
|
|
delete_channel(channel);
|
|
return;
|
|
}
|
|
|
|
/* No peer_fd means a subd crash or disconnection. */
|
|
if (!peer_fd) {
|
|
/* If the channel is unsaved, we forget it */
|
|
channel_fail_reconnect(channel, "%s: %s",
|
|
channel->owner->name, desc);
|
|
return;
|
|
}
|
|
|
|
/* Do we have an error to send? */
|
|
if (err_for_them && !channel->error && !warning)
|
|
channel->error = tal_dup_talarr(channel, u8, err_for_them);
|
|
|
|
/* Other implementations chose to ignore errors early on. Not
|
|
* surprisingly, they now spew out spurious errors frequently,
|
|
* and we would close the channel on them. We now support warnings
|
|
* for this case. */
|
|
if (warning) {
|
|
channel_fail_reconnect_later(channel, "%s WARNING: %s",
|
|
channel->owner->name, desc);
|
|
return;
|
|
}
|
|
|
|
/* BOLT #1:
|
|
*
|
|
* A sending node:
|
|
*...
|
|
* - when `channel_id` is 0:
|
|
* - MUST fail all channels with the receiving node.
|
|
* - MUST close the connection.
|
|
*/
|
|
/* FIXME: Close if it's an all-channels error sent or rcvd */
|
|
|
|
/* BOLT #1:
|
|
*
|
|
* A sending node:
|
|
* - when sending `error`:
|
|
* - MUST fail the channel referred to by the error message.
|
|
*...
|
|
* The receiving node:
|
|
* - upon receiving `error`:
|
|
* - MUST fail the channel referred to by the error message,
|
|
* if that channel is with the sending node.
|
|
*/
|
|
|
|
/* We should immediately forget the channel if we receive error during
|
|
* CHANNELD_AWAITING_LOCKIN if we are fundee. */
|
|
if (!err_for_them && channel->opener == REMOTE
|
|
&& channel->state == CHANNELD_AWAITING_LOCKIN)
|
|
channel_fail_forget(channel, "%s: %s ERROR %s",
|
|
channel->owner->name,
|
|
err_for_them ? "sent" : "received", desc);
|
|
else
|
|
channel_fail_permanent(channel,
|
|
err_for_them ? REASON_LOCAL : REASON_PROTOCOL,
|
|
"%s: %s ERROR %s",
|
|
channel->owner->name,
|
|
err_for_them ? "sent" : "received", desc);
|
|
}
|
|
|
|
static void json_add_htlcs(struct lightningd *ld,
|
|
struct json_stream *response,
|
|
const struct channel *channel)
|
|
{
|
|
/* FIXME: make per-channel htlc maps! */
|
|
const struct htlc_in *hin;
|
|
struct htlc_in_map_iter ini;
|
|
const struct htlc_out *hout;
|
|
struct htlc_out_map_iter outi;
|
|
u32 local_feerate = get_feerate(channel->fee_states,
|
|
channel->opener, LOCAL);
|
|
|
|
/* FIXME: Add more fields. */
|
|
json_array_start(response, "htlcs");
|
|
for (hin = htlc_in_map_first(&ld->htlcs_in, &ini);
|
|
hin;
|
|
hin = htlc_in_map_next(&ld->htlcs_in, &ini)) {
|
|
if (hin->key.channel != channel)
|
|
continue;
|
|
|
|
json_object_start(response, NULL);
|
|
json_add_string(response, "direction", "in");
|
|
json_add_u64(response, "id", hin->key.id);
|
|
json_add_amount_msat_compat(response, hin->msat,
|
|
"msatoshi", "amount_msat");
|
|
json_add_u32(response, "expiry", hin->cltv_expiry);
|
|
json_add_sha256(response, "payment_hash", &hin->payment_hash);
|
|
json_add_string(response, "state",
|
|
htlc_state_name(hin->hstate));
|
|
if (htlc_is_trimmed(REMOTE, hin->msat, local_feerate,
|
|
channel->our_config.dust_limit, LOCAL,
|
|
channel_has(channel, OPT_ANCHOR_OUTPUTS)))
|
|
json_add_bool(response, "local_trimmed", true);
|
|
if (hin->status != NULL)
|
|
json_add_string(response, "status", hin->status);
|
|
json_object_end(response);
|
|
}
|
|
|
|
for (hout = htlc_out_map_first(&ld->htlcs_out, &outi);
|
|
hout;
|
|
hout = htlc_out_map_next(&ld->htlcs_out, &outi)) {
|
|
if (hout->key.channel != channel)
|
|
continue;
|
|
|
|
json_object_start(response, NULL);
|
|
json_add_string(response, "direction", "out");
|
|
json_add_u64(response, "id", hout->key.id);
|
|
json_add_amount_msat_compat(response, hout->msat,
|
|
"msatoshi", "amount_msat");
|
|
json_add_u64(response, "expiry", hout->cltv_expiry);
|
|
json_add_sha256(response, "payment_hash", &hout->payment_hash);
|
|
json_add_string(response, "state",
|
|
htlc_state_name(hout->hstate));
|
|
if (htlc_is_trimmed(LOCAL, hout->msat, local_feerate,
|
|
channel->our_config.dust_limit, LOCAL,
|
|
channel_has(channel, OPT_ANCHOR_OUTPUTS)))
|
|
json_add_bool(response, "local_trimmed", true);
|
|
json_object_end(response);
|
|
}
|
|
json_array_end(response);
|
|
}
|
|
|
|
/* We do this replication manually because it's an array. */
|
|
static void json_add_sat_only(struct json_stream *result,
|
|
const char *fieldname,
|
|
struct amount_sat sat)
|
|
{
|
|
struct amount_msat msat;
|
|
|
|
if (amount_sat_to_msat(&msat, sat))
|
|
json_add_string(result, fieldname,
|
|
type_to_string(tmpctx, struct amount_msat, &msat));
|
|
}
|
|
|
|
/* Fee a commitment transaction would currently cost */
|
|
static struct amount_sat commit_txfee(const struct channel *channel,
|
|
struct amount_msat amount,
|
|
enum side side)
|
|
{
|
|
/* FIXME: make per-channel htlc maps! */
|
|
const struct htlc_in *hin;
|
|
struct htlc_in_map_iter ini;
|
|
const struct htlc_out *hout;
|
|
struct htlc_out_map_iter outi;
|
|
struct lightningd *ld = channel->peer->ld;
|
|
size_t num_untrimmed_htlcs = 0;
|
|
u32 feerate = get_feerate(channel->fee_states,
|
|
channel->opener, side);
|
|
struct amount_sat dust_limit;
|
|
struct amount_sat fee;
|
|
bool option_anchor_outputs = channel_has(channel, OPT_ANCHOR_OUTPUTS);
|
|
|
|
if (side == LOCAL)
|
|
dust_limit = channel->our_config.dust_limit;
|
|
if (side == REMOTE)
|
|
dust_limit = channel->channel_info.their_config.dust_limit;
|
|
|
|
/* Assume we tried to add "amount" */
|
|
if (!htlc_is_trimmed(side, amount, feerate, dust_limit, side,
|
|
option_anchor_outputs))
|
|
num_untrimmed_htlcs++;
|
|
|
|
for (hin = htlc_in_map_first(&ld->htlcs_in, &ini);
|
|
hin;
|
|
hin = htlc_in_map_next(&ld->htlcs_in, &ini)) {
|
|
if (hin->key.channel != channel)
|
|
continue;
|
|
if (!htlc_is_trimmed(!side, hin->msat, feerate, dust_limit,
|
|
side, option_anchor_outputs))
|
|
num_untrimmed_htlcs++;
|
|
}
|
|
for (hout = htlc_out_map_first(&ld->htlcs_out, &outi);
|
|
hout;
|
|
hout = htlc_out_map_next(&ld->htlcs_out, &outi)) {
|
|
if (hout->key.channel != channel)
|
|
continue;
|
|
if (!htlc_is_trimmed(side, hout->msat, feerate, dust_limit,
|
|
side, option_anchor_outputs))
|
|
num_untrimmed_htlcs++;
|
|
}
|
|
|
|
/*
|
|
* BOLT #2:
|
|
* A sending node:
|
|
*...
|
|
* - SHOULD NOT offer `amount_msat` if, after adding that HTLC to its
|
|
* commitment transaction, its remaining balance doesn't allow it to
|
|
* pay the commitment transaction fee when receiving or sending a
|
|
* future additional non-dust HTLC while maintaining its channel
|
|
* reserve. It is recommended that this "fee spike buffer" can
|
|
* handle twice the current `feerate_per_kw` to ensure
|
|
* predictability between implementations.
|
|
*/
|
|
fee = commit_tx_base_fee(2 * feerate, num_untrimmed_htlcs + 1,
|
|
option_anchor_outputs);
|
|
|
|
if (option_anchor_outputs) {
|
|
/* BOLT #3:
|
|
* If `option_anchors` applies to the commitment
|
|
* transaction, also subtract two times the fixed anchor size
|
|
* of 330 sats from the funder (either `to_local` or
|
|
* `to_remote`).
|
|
*/
|
|
if (!amount_sat_add(&fee, fee, AMOUNT_SAT(660)))
|
|
; /* fee is somehow astronomical already.... */
|
|
}
|
|
|
|
return fee;
|
|
}
|
|
|
|
static void subtract_offered_htlcs(const struct channel *channel,
|
|
struct amount_msat *amount)
|
|
{
|
|
const struct htlc_out *hout;
|
|
struct htlc_out_map_iter outi;
|
|
struct lightningd *ld = channel->peer->ld;
|
|
|
|
for (hout = htlc_out_map_first(&ld->htlcs_out, &outi);
|
|
hout;
|
|
hout = htlc_out_map_next(&ld->htlcs_out, &outi)) {
|
|
if (hout->key.channel != channel)
|
|
continue;
|
|
if (!amount_msat_sub(amount, *amount, hout->msat))
|
|
*amount = AMOUNT_MSAT(0);
|
|
}
|
|
}
|
|
|
|
static void subtract_received_htlcs(const struct channel *channel,
|
|
struct amount_msat *amount)
|
|
{
|
|
const struct htlc_in *hin;
|
|
struct htlc_in_map_iter ini;
|
|
struct lightningd *ld = channel->peer->ld;
|
|
|
|
for (hin = htlc_in_map_first(&ld->htlcs_in, &ini);
|
|
hin;
|
|
hin = htlc_in_map_next(&ld->htlcs_in, &ini)) {
|
|
if (hin->key.channel != channel)
|
|
continue;
|
|
if (!amount_msat_sub(amount, *amount, hin->msat))
|
|
*amount = AMOUNT_MSAT(0);
|
|
}
|
|
}
|
|
|
|
struct amount_msat channel_amount_spendable(const struct channel *channel)
|
|
{
|
|
struct amount_msat spendable;
|
|
|
|
/* Compute how much we can send via this channel in one payment. */
|
|
if (!amount_msat_sub_sat(&spendable,
|
|
channel->our_msat,
|
|
channel->channel_info.their_config.channel_reserve))
|
|
return AMOUNT_MSAT(0);
|
|
|
|
/* Take away any currently-offered HTLCs. */
|
|
subtract_offered_htlcs(channel, &spendable);
|
|
|
|
/* If we're opener, subtract txfees we'll need to spend this */
|
|
if (channel->opener == LOCAL) {
|
|
if (!amount_msat_sub_sat(&spendable, spendable,
|
|
commit_txfee(channel, spendable,
|
|
LOCAL)))
|
|
return AMOUNT_MSAT(0);
|
|
}
|
|
|
|
/* We can't offer an HTLC less than the other side will accept. */
|
|
if (amount_msat_less(spendable,
|
|
channel->channel_info.their_config.htlc_minimum))
|
|
return AMOUNT_MSAT(0);
|
|
|
|
/* We can't offer an HTLC over the max payment threshold either. */
|
|
if (amount_msat_greater(spendable, chainparams->max_payment))
|
|
spendable = chainparams->max_payment;
|
|
|
|
return spendable;
|
|
}
|
|
|
|
struct amount_msat channel_amount_receivable(const struct channel *channel)
|
|
{
|
|
struct amount_msat their_msat, receivable;
|
|
|
|
/* Compute how much we can receive via this channel in one payment */
|
|
if (!amount_sat_sub_msat(&their_msat,
|
|
channel->funding_sats, channel->our_msat))
|
|
their_msat = AMOUNT_MSAT(0);
|
|
|
|
if (!amount_msat_sub_sat(&receivable,
|
|
their_msat,
|
|
channel->our_config.channel_reserve))
|
|
return AMOUNT_MSAT(0);
|
|
|
|
/* Take away any currently-offered HTLCs. */
|
|
subtract_received_htlcs(channel, &receivable);
|
|
|
|
/* If they're opener, subtract txfees they'll need to spend this */
|
|
if (channel->opener == REMOTE) {
|
|
if (!amount_msat_sub_sat(&receivable, receivable,
|
|
commit_txfee(channel,
|
|
receivable, REMOTE)))
|
|
return AMOUNT_MSAT(0);
|
|
}
|
|
|
|
/* They can't offer an HTLC less than what we will accept. */
|
|
if (amount_msat_less(receivable, channel->our_config.htlc_minimum))
|
|
return AMOUNT_MSAT(0);
|
|
|
|
/* They can't offer an HTLC over the max payment threshold either. */
|
|
if (amount_msat_greater(receivable, chainparams->max_payment))
|
|
receivable = chainparams->max_payment;
|
|
|
|
return receivable;
|
|
}
|
|
|
|
static void json_add_channel(struct lightningd *ld,
|
|
struct json_stream *response, const char *key,
|
|
const struct channel *channel)
|
|
{
|
|
struct channel_stats channel_stats;
|
|
struct amount_msat funding_msat, peer_msats, our_msats;
|
|
struct amount_sat peer_funded_sats;
|
|
struct peer *p = channel->peer;
|
|
struct state_change_entry *state_changes;
|
|
u32 feerate;
|
|
|
|
json_object_start(response, key);
|
|
json_add_string(response, "state", channel_state_name(channel));
|
|
if (channel->last_tx && !invalid_last_tx(channel->last_tx)) {
|
|
struct bitcoin_txid txid;
|
|
bitcoin_txid(channel->last_tx, &txid);
|
|
|
|
json_add_txid(response, "scratch_txid", &txid);
|
|
if (deprecated_apis)
|
|
json_add_amount_sat_only(response, "last_tx_fee",
|
|
bitcoin_tx_compute_fee(channel->last_tx));
|
|
json_add_amount_sat_only(response, "last_tx_fee_msat",
|
|
bitcoin_tx_compute_fee(channel->last_tx));
|
|
}
|
|
|
|
json_object_start(response, "feerate");
|
|
feerate = get_feerate(channel->fee_states, channel->opener, LOCAL);
|
|
json_add_u32(response, feerate_style_name(FEERATE_PER_KSIPA), feerate);
|
|
json_add_u32(response, feerate_style_name(FEERATE_PER_KBYTE),
|
|
feerate_to_style(feerate, FEERATE_PER_KBYTE));
|
|
json_object_end(response);
|
|
|
|
if (channel->owner)
|
|
json_add_string(response, "owner", channel->owner->name);
|
|
|
|
if (channel->scid) {
|
|
json_add_short_channel_id(response, "short_channel_id",
|
|
channel->scid);
|
|
json_add_num(response, "direction",
|
|
node_id_idx(&ld->id, &channel->peer->id));
|
|
}
|
|
|
|
json_add_string(response, "channel_id",
|
|
type_to_string(tmpctx, struct channel_id, &channel->cid));
|
|
json_add_txid(response, "funding_txid", &channel->funding.txid);
|
|
json_add_num(response, "funding_outnum", channel->funding.n);
|
|
|
|
if (!list_empty(&channel->inflights)) {
|
|
struct channel_inflight *initial, *inflight;
|
|
u32 last_feerate, next_feerate;
|
|
|
|
initial = list_top(&channel->inflights,
|
|
struct channel_inflight, list);
|
|
json_add_string(response, "initial_feerate",
|
|
tal_fmt(tmpctx, "%d%s",
|
|
initial->funding->feerate,
|
|
feerate_style_name(FEERATE_PER_KSIPA)));
|
|
|
|
last_feerate = channel_last_funding_feerate(channel);
|
|
assert(last_feerate > 0);
|
|
json_add_string(response, "last_feerate",
|
|
tal_fmt(tmpctx, "%d%s", last_feerate,
|
|
feerate_style_name(FEERATE_PER_KSIPA)));
|
|
|
|
/* BOLT-9e7723387c8859b511e178485605a0b9133b9869 #2:
|
|
* - MUST set `funding_feerate_perkw` greater than or equal to
|
|
* 65/64 times the last sent `funding_feerate_perkw`
|
|
* rounded down.
|
|
*/
|
|
next_feerate = last_feerate * 65 / 64;
|
|
assert(next_feerate > last_feerate);
|
|
json_add_string(response, "next_feerate",
|
|
tal_fmt(tmpctx, "%d%s", next_feerate,
|
|
feerate_style_name(FEERATE_PER_KSIPA)));
|
|
|
|
/* List the inflights */
|
|
json_array_start(response, "inflight");
|
|
list_for_each(&channel->inflights, inflight, list) {
|
|
struct bitcoin_txid txid;
|
|
|
|
json_object_start(response, NULL);
|
|
json_add_txid(response, "funding_txid",
|
|
&inflight->funding->outpoint.txid);
|
|
json_add_num(response, "funding_outnum",
|
|
inflight->funding->outpoint.n);
|
|
json_add_string(response, "feerate",
|
|
tal_fmt(tmpctx, "%d%s",
|
|
inflight->funding->feerate,
|
|
feerate_style_name(
|
|
FEERATE_PER_KSIPA)));
|
|
json_add_amount_sat_only(response,
|
|
"total_funding_msat",
|
|
inflight->funding->total_funds);
|
|
json_add_amount_sat_only(response,
|
|
"our_funding_msat",
|
|
inflight->funding->our_funds);
|
|
/* Add the expected commitment tx id also */
|
|
bitcoin_txid(inflight->last_tx, &txid);
|
|
json_add_txid(response, "scratch_txid", &txid);
|
|
json_object_end(response);
|
|
}
|
|
json_array_end(response);
|
|
}
|
|
|
|
if (channel->shutdown_scriptpubkey[LOCAL]) {
|
|
char *addr = encode_scriptpubkey_to_addr(tmpctx,
|
|
chainparams,
|
|
channel->shutdown_scriptpubkey[LOCAL]);
|
|
if (addr)
|
|
json_add_string(response, "close_to_addr", addr);
|
|
json_add_hex_talarr(response, "close_to",
|
|
channel->shutdown_scriptpubkey[LOCAL]);
|
|
}
|
|
|
|
json_add_bool(
|
|
response, "private",
|
|
!(channel->channel_flags & CHANNEL_FLAGS_ANNOUNCE_CHANNEL));
|
|
|
|
/* opener and closer */
|
|
assert(channel->opener != NUM_SIDES);
|
|
json_add_string(response, "opener", channel->opener == LOCAL ?
|
|
"local" : "remote");
|
|
if (channel->closer != NUM_SIDES)
|
|
json_add_string(response, "closer", channel->closer == LOCAL ?
|
|
"local" : "remote");
|
|
else if (deprecated_apis)
|
|
json_add_null(response, "closer");
|
|
|
|
json_array_start(response, "features");
|
|
if (channel_has(channel, OPT_STATIC_REMOTEKEY))
|
|
json_add_string(response, NULL, "option_static_remotekey");
|
|
if (channel_has(channel, OPT_ANCHOR_OUTPUTS))
|
|
json_add_string(response, NULL, "option_anchor_outputs");
|
|
json_array_end(response);
|
|
|
|
if (!amount_sat_sub(&peer_funded_sats, channel->funding_sats,
|
|
channel->our_funds)) {
|
|
log_broken(channel->log,
|
|
"Overflow subtracing funding %s, our funds %s",
|
|
type_to_string(tmpctx, struct amount_sat,
|
|
&channel->funding_sats),
|
|
type_to_string(tmpctx, struct amount_sat,
|
|
&channel->our_funds));
|
|
peer_funded_sats = AMOUNT_SAT(0);
|
|
}
|
|
if (!amount_sat_to_msat(&peer_msats, peer_funded_sats)) {
|
|
log_broken(channel->log,
|
|
"Overflow converting peer sats %s to msat",
|
|
type_to_string(tmpctx, struct amount_sat,
|
|
&peer_funded_sats));
|
|
peer_msats = AMOUNT_MSAT(0);
|
|
}
|
|
if (!amount_sat_to_msat(&our_msats, channel->our_funds)) {
|
|
log_broken(channel->log,
|
|
"Overflow converting peer sats %s to msat",
|
|
type_to_string(tmpctx, struct amount_sat,
|
|
&channel->our_funds));
|
|
our_msats = AMOUNT_MSAT(0);
|
|
}
|
|
|
|
if (deprecated_apis) {
|
|
json_object_start(response, "funding_allocation_msat");
|
|
json_add_u64(response, node_id_to_hexstr(tmpctx, &p->id),
|
|
peer_msats.millisatoshis); /* Raw: JSON field */
|
|
json_add_u64(response, node_id_to_hexstr(tmpctx, &ld->id),
|
|
our_msats.millisatoshis); /* Raw: JSON field */
|
|
json_object_end(response);
|
|
|
|
json_object_start(response, "funding_msat");
|
|
json_add_sat_only(response,
|
|
node_id_to_hexstr(tmpctx, &p->id),
|
|
peer_funded_sats);
|
|
json_add_sat_only(response,
|
|
node_id_to_hexstr(tmpctx, &ld->id),
|
|
channel->our_funds);
|
|
json_object_end(response);
|
|
}
|
|
|
|
json_object_start(response, "funding");
|
|
json_add_sat_only(response, "local_msat", channel->our_funds);
|
|
json_add_sat_only(response, "remote_msat", peer_funded_sats);
|
|
json_add_amount_msat_only(response, "pushed_msat", channel->push);
|
|
json_object_end(response);
|
|
|
|
if (!amount_sat_to_msat(&funding_msat, channel->funding_sats)) {
|
|
log_broken(channel->log,
|
|
"Overflow converting funding %s",
|
|
type_to_string(tmpctx, struct amount_sat,
|
|
&channel->funding_sats));
|
|
funding_msat = AMOUNT_MSAT(0);
|
|
}
|
|
json_add_amount_msat_compat(response, channel->our_msat,
|
|
"msatoshi_to_us", "to_us_msat");
|
|
json_add_amount_msat_compat(response, channel->msat_to_us_min,
|
|
"msatoshi_to_us_min", "min_to_us_msat");
|
|
json_add_amount_msat_compat(response, channel->msat_to_us_max,
|
|
"msatoshi_to_us_max", "max_to_us_msat");
|
|
json_add_amount_msat_compat(response, funding_msat,
|
|
"msatoshi_total", "total_msat");
|
|
|
|
/* routing fees */
|
|
json_add_amount_msat_only(response, "fee_base_msat",
|
|
amount_msat(channel->feerate_base));
|
|
json_add_u32(response, "fee_proportional_millionths",
|
|
channel->feerate_ppm);
|
|
|
|
/* channel config */
|
|
json_add_amount_sat_compat(response,
|
|
channel->our_config.dust_limit,
|
|
"dust_limit_satoshis", "dust_limit_msat");
|
|
json_add_amount_msat_compat(response,
|
|
channel->our_config.max_htlc_value_in_flight,
|
|
"max_htlc_value_in_flight_msat",
|
|
"max_total_htlc_in_msat");
|
|
|
|
/* The `channel_reserve_satoshis` is imposed on
|
|
* the *other* side (see `channel_reserve_msat`
|
|
* function in, it uses `!side` to flip sides).
|
|
* So our configuration `channel_reserve_satoshis`
|
|
* is imposed on their side, while their
|
|
* configuration `channel_reserve_satoshis` is
|
|
* imposed on ours. */
|
|
json_add_amount_sat_compat(response,
|
|
channel->our_config.channel_reserve,
|
|
"their_channel_reserve_satoshis",
|
|
"their_reserve_msat");
|
|
json_add_amount_sat_compat(response,
|
|
channel->channel_info.their_config.channel_reserve,
|
|
"our_channel_reserve_satoshis",
|
|
"our_reserve_msat");
|
|
|
|
/* append spendable to JSON output */
|
|
json_add_amount_msat_compat(response,
|
|
channel_amount_spendable(channel),
|
|
"spendable_msatoshi", "spendable_msat");
|
|
|
|
/* append receivable to JSON output */
|
|
json_add_amount_msat_compat(response,
|
|
channel_amount_receivable(channel),
|
|
"receivable_msatoshi", "receivable_msat");
|
|
|
|
json_add_amount_msat_compat(response,
|
|
channel->our_config.htlc_minimum,
|
|
"htlc_minimum_msat",
|
|
"minimum_htlc_in_msat");
|
|
json_add_amount_msat_only(response,
|
|
"minimum_htlc_out_msat",
|
|
channel->htlc_minimum_msat);
|
|
json_add_amount_msat_only(response,
|
|
"maximum_htlc_out_msat",
|
|
channel->htlc_maximum_msat);
|
|
|
|
/* The `to_self_delay` is imposed on the *other*
|
|
* side, so our configuration `to_self_delay` is
|
|
* imposed on their side, while their configuration
|
|
* `to_self_delay` is imposed on ours. */
|
|
json_add_num(response, "their_to_self_delay",
|
|
channel->our_config.to_self_delay);
|
|
json_add_num(response, "our_to_self_delay",
|
|
channel->channel_info.their_config.to_self_delay);
|
|
json_add_num(response, "max_accepted_htlcs",
|
|
channel->our_config.max_accepted_htlcs);
|
|
|
|
state_changes = wallet_state_change_get(ld->wallet, tmpctx, channel->dbid);
|
|
json_array_start(response, "state_changes");
|
|
for (size_t i = 0; i < tal_count(state_changes); i++) {
|
|
json_object_start(response, NULL);
|
|
json_add_timeiso(response, "timestamp",
|
|
&state_changes[i].timestamp);
|
|
json_add_string(response, "old_state",
|
|
channel_state_str(state_changes[i].old_state));
|
|
json_add_string(response, "new_state",
|
|
channel_state_str(state_changes[i].new_state));
|
|
json_add_string(response, "cause",
|
|
channel_change_state_reason_str(state_changes[i].cause));
|
|
json_add_string(response, "message", state_changes[i].message);
|
|
json_object_end(response);
|
|
}
|
|
json_array_end(response);
|
|
|
|
json_array_start(response, "status");
|
|
for (size_t i = 0; i < ARRAY_SIZE(channel->billboard.permanent); i++) {
|
|
if (!channel->billboard.permanent[i])
|
|
continue;
|
|
json_add_string(response, NULL,
|
|
channel->billboard.permanent[i]);
|
|
}
|
|
if (channel->billboard.transient)
|
|
json_add_string(response, NULL, channel->billboard.transient);
|
|
json_array_end(response);
|
|
|
|
/* Provide channel statistics */
|
|
wallet_channel_stats_load(ld->wallet, channel->dbid, &channel_stats);
|
|
json_add_u64(response, "in_payments_offered",
|
|
channel_stats.in_payments_offered);
|
|
json_add_amount_msat_compat(response,
|
|
channel_stats.in_msatoshi_offered,
|
|
"in_msatoshi_offered",
|
|
"in_offered_msat");
|
|
json_add_u64(response, "in_payments_fulfilled",
|
|
channel_stats.in_payments_fulfilled);
|
|
json_add_amount_msat_compat(response,
|
|
channel_stats.in_msatoshi_fulfilled,
|
|
"in_msatoshi_fulfilled",
|
|
"in_fulfilled_msat");
|
|
json_add_u64(response, "out_payments_offered",
|
|
channel_stats.out_payments_offered);
|
|
json_add_amount_msat_compat(response,
|
|
channel_stats.out_msatoshi_offered,
|
|
"out_msatoshi_offered",
|
|
"out_offered_msat");
|
|
json_add_u64(response, "out_payments_fulfilled",
|
|
channel_stats.out_payments_fulfilled);
|
|
json_add_amount_msat_compat(response,
|
|
channel_stats.out_msatoshi_fulfilled,
|
|
"out_msatoshi_fulfilled",
|
|
"out_fulfilled_msat");
|
|
|
|
json_add_htlcs(ld, response, channel);
|
|
json_object_end(response);
|
|
}
|
|
|
|
struct peer_connected_hook_payload {
|
|
struct lightningd *ld;
|
|
struct wireaddr_internal addr;
|
|
struct wireaddr *remote_addr;
|
|
bool incoming;
|
|
struct peer *peer;
|
|
u8 *error;
|
|
};
|
|
|
|
static void
|
|
peer_connected_serialize(struct peer_connected_hook_payload *payload,
|
|
struct json_stream *stream, struct plugin *plugin)
|
|
{
|
|
const struct peer *p = payload->peer;
|
|
json_object_start(stream, "peer");
|
|
json_add_node_id(stream, "id", &p->id);
|
|
json_add_string(stream, "direction", payload->incoming ? "in" : "out");
|
|
json_add_string(
|
|
stream, "addr",
|
|
type_to_string(stream, struct wireaddr_internal, &payload->addr));
|
|
if (payload->remote_addr)
|
|
json_add_string(
|
|
stream, "remote_addr",
|
|
type_to_string(stream, struct wireaddr, payload->remote_addr));
|
|
json_add_hex_talarr(stream, "features", p->their_features);
|
|
json_object_end(stream); /* .peer */
|
|
}
|
|
|
|
static void peer_connected_hook_final(struct peer_connected_hook_payload *payload STEALS)
|
|
{
|
|
struct lightningd *ld = payload->ld;
|
|
struct channel *channel;
|
|
struct wireaddr_internal addr = payload->addr;
|
|
struct peer *peer = payload->peer;
|
|
u8 *error;
|
|
|
|
/* Whatever happens, we free payload (it's currently a child
|
|
* of the peer, which may be freed if we fail to start
|
|
* subd). */
|
|
tal_steal(tmpctx, payload);
|
|
|
|
/* Notify anyone who cares */
|
|
notify_connect(ld, &peer->id, payload->incoming, &addr);
|
|
|
|
/* Check for specific errors of a hook */
|
|
if (payload->error) {
|
|
error = payload->error;
|
|
goto send_error;
|
|
}
|
|
|
|
list_for_each(&peer->channels, channel, list) {
|
|
#if DEVELOPER
|
|
if (dev_disconnect_permanent(ld)) {
|
|
channel_fail_permanent(channel, REASON_LOCAL,
|
|
"dev_disconnect permfail");
|
|
error = channel->error;
|
|
goto send_error;
|
|
}
|
|
#endif
|
|
|
|
if (channel_active(channel)) {
|
|
log_debug(channel->log, "Peer has reconnected, state %s: telling connectd to make active",
|
|
channel_state_name(channel));
|
|
|
|
assert(!channel->owner);
|
|
subd_send_msg(ld->connectd,
|
|
take(towire_connectd_peer_make_active(NULL, &peer->id,
|
|
&channel->cid)));
|
|
}
|
|
}
|
|
return;
|
|
|
|
send_error:
|
|
log_peer_debug(ld->log, &peer->id, "Telling connectd to send error %s",
|
|
tal_hex(tmpctx, error));
|
|
/* Get connectd to send error and close. */
|
|
subd_send_msg(ld->connectd,
|
|
take(towire_connectd_peer_final_msg(NULL, &peer->id,
|
|
error)));
|
|
}
|
|
|
|
static bool
|
|
peer_connected_hook_deserialize(struct peer_connected_hook_payload *payload,
|
|
const char *buffer,
|
|
const jsmntok_t *toks)
|
|
{
|
|
struct lightningd *ld = payload->ld;
|
|
|
|
/* already rejected by prior plugin hook in the chain */
|
|
if (payload->error != NULL)
|
|
return true;
|
|
|
|
if (!toks || !buffer)
|
|
return true;
|
|
|
|
/* If we had a hook, interpret result. */
|
|
const jsmntok_t *t_res = json_get_member(buffer, toks, "result");
|
|
const jsmntok_t *t_err = json_get_member(buffer, toks, "error_message");
|
|
|
|
/* fail */
|
|
if (!t_res)
|
|
fatal("Plugin returned an invalid response to the "
|
|
"peer_connected hook: %s", buffer);
|
|
|
|
/* reject */
|
|
if (json_tok_streq(buffer, t_res, "disconnect")) {
|
|
payload->error = (u8*)"";
|
|
if (t_err) {
|
|
payload->error = towire_warningfmt(tmpctx, NULL, "%.*s",
|
|
t_err->end - t_err->start,
|
|
buffer + t_err->start);
|
|
}
|
|
log_debug(ld->log, "peer_connected hook rejects and says '%s'",
|
|
payload->error);
|
|
/* At this point we suppress other plugins in the chain and
|
|
* directly move to final */
|
|
peer_connected_hook_final(payload);
|
|
return false;
|
|
} else if (!json_tok_streq(buffer, t_res, "continue"))
|
|
fatal("Plugin returned an invalid response to the "
|
|
"peer_connected hook: %s", buffer);
|
|
|
|
/* call next hook */
|
|
return true;
|
|
}
|
|
|
|
/* Compare and store `remote_addr` and the `peer_id` that reported it.
|
|
* If new address was reported by at least one other, do node_announcement */
|
|
static void update_remote_addr(struct lightningd *ld,
|
|
const struct wireaddr *remote_addr,
|
|
const struct node_id peer_id)
|
|
{
|
|
/* failsafe to prevent privacy leakage. */
|
|
if (ld->always_use_proxy)
|
|
return;
|
|
|
|
switch (remote_addr->type) {
|
|
case ADDR_TYPE_IPV4:
|
|
/* init pointers first time */
|
|
if (ld->remote_addr_v4 == NULL) {
|
|
ld->remote_addr_v4 = tal_dup(ld, struct wireaddr,
|
|
remote_addr);
|
|
ld->remote_addr_v4_peer = peer_id;
|
|
}
|
|
/* if updated by the same peer just remember the latest addr */
|
|
if (node_id_eq(&ld->remote_addr_v4_peer, &peer_id)) {
|
|
*ld->remote_addr_v4 = *remote_addr;
|
|
break;
|
|
}
|
|
/* tell gossip we have a valid update */
|
|
if (wireaddr_eq_without_port(ld->remote_addr_v4, remote_addr))
|
|
subd_send_msg(ld->gossip, towire_gossipd_remote_addr(
|
|
tmpctx,
|
|
ld->remote_addr_v4));
|
|
/* store latest values */
|
|
*ld->remote_addr_v4 = *remote_addr;
|
|
ld->remote_addr_v4_peer = peer_id;
|
|
break;
|
|
case ADDR_TYPE_IPV6:
|
|
/* same code :s/4/6/ without the comments ;) */
|
|
if (ld->remote_addr_v6 == NULL) {
|
|
ld->remote_addr_v6 = tal_dup(ld, struct wireaddr,
|
|
remote_addr);
|
|
ld->remote_addr_v6_peer = peer_id;
|
|
}
|
|
if (node_id_eq(&ld->remote_addr_v6_peer, &peer_id)) {
|
|
*ld->remote_addr_v6 = *remote_addr;
|
|
break;
|
|
}
|
|
if (wireaddr_eq_without_port(ld->remote_addr_v6, remote_addr))
|
|
subd_send_msg(ld->gossip, towire_gossipd_remote_addr(
|
|
tmpctx,
|
|
ld->remote_addr_v6));
|
|
*ld->remote_addr_v6 = *remote_addr;
|
|
ld->remote_addr_v6_peer = peer_id;
|
|
break;
|
|
|
|
/* ignore all other cases */
|
|
case ADDR_TYPE_TOR_V2_REMOVED:
|
|
case ADDR_TYPE_TOR_V3:
|
|
case ADDR_TYPE_DNS:
|
|
case ADDR_TYPE_WEBSOCKET:
|
|
break;
|
|
}
|
|
}
|
|
|
|
REGISTER_PLUGIN_HOOK(peer_connected,
|
|
peer_connected_hook_deserialize,
|
|
peer_connected_hook_final,
|
|
peer_connected_serialize,
|
|
struct peer_connected_hook_payload *);
|
|
|
|
/* Returns true if we're still waiting for subds for active channels */
|
|
static bool peer_subds_pending(const struct peer *peer)
|
|
{
|
|
struct channel *channel;
|
|
|
|
list_for_each(&peer->channels, channel, list) {
|
|
if (!channel_active(channel))
|
|
continue;
|
|
if (!channel->owner)
|
|
return true;
|
|
}
|
|
return false;
|
|
}
|
|
|
|
/* Connectd tells us a peer has connected: it never hands us duplicates, since
|
|
* it holds them until we say peer_disconnected. */
|
|
void peer_connected(struct lightningd *ld, const u8 *msg)
|
|
{
|
|
struct node_id id;
|
|
u8 *their_features;
|
|
struct peer *peer;
|
|
struct peer_connected_hook_payload *hook_payload;
|
|
|
|
hook_payload = tal(NULL, struct peer_connected_hook_payload);
|
|
hook_payload->ld = ld;
|
|
hook_payload->error = NULL;
|
|
if (!fromwire_connectd_peer_connected(hook_payload, msg,
|
|
&id, &hook_payload->addr,
|
|
&hook_payload->remote_addr,
|
|
&hook_payload->incoming,
|
|
&their_features))
|
|
fatal("Connectd gave bad CONNECT_PEER_CONNECTED message %s",
|
|
tal_hex(msg, msg));
|
|
|
|
/* If we're already dealing with this peer, hand off to correct
|
|
* subdaemon. Otherwise, we'll hand to openingd to wait there. */
|
|
peer = peer_by_id(ld, &id);
|
|
if (!peer)
|
|
peer = new_peer(ld, 0, &id, &hook_payload->addr,
|
|
hook_payload->incoming);
|
|
peer->is_connected = true;
|
|
/* Update peer address and direction */
|
|
peer->addr = hook_payload->addr;
|
|
peer->connected_incoming = hook_payload->incoming;
|
|
peer_update_features(peer, their_features);
|
|
|
|
tal_steal(peer, hook_payload);
|
|
hook_payload->peer = peer;
|
|
|
|
/* Complete any outstanding connect commands: as a hack, we delay here if we
|
|
* are going to make them active (so when connect returns, the channels are ready).
|
|
* So we also wake these up if the connection dies before that! */
|
|
if (!peer_subds_pending(peer))
|
|
connect_succeeded(ld, peer, hook_payload->incoming, &hook_payload->addr);
|
|
|
|
/* Can't be opening, since we wouldn't have sent peer_disconnected. */
|
|
assert(!peer->uncommitted_channel);
|
|
|
|
/* Log and update remote_addr for Nat/IP discovery. */
|
|
if (hook_payload->remote_addr) {
|
|
log_info(ld->log, "Peer says it sees our address as: %s",
|
|
fmt_wireaddr(tmpctx, hook_payload->remote_addr));
|
|
/* Currently only from peers we have a channel with, until we
|
|
* do stuff like probing for remote_addr to a random node. */
|
|
if (!list_empty(&peer->channels))
|
|
update_remote_addr(ld, hook_payload->remote_addr, id);
|
|
}
|
|
|
|
plugin_hook_call_peer_connected(ld, hook_payload);
|
|
}
|
|
|
|
/* connectd tells us a peer has an interesting message, and hands us an
|
|
* fd to give to the correct subdaemon. Unlike peer_connected, this is racy:
|
|
* we might have just told it to disconnect peer. */
|
|
void peer_active(struct lightningd *ld, const u8 *msg, int fd)
|
|
{
|
|
struct node_id id;
|
|
u16 *msgtype;
|
|
struct channel *channel;
|
|
struct channel_id channel_id;
|
|
struct peer *peer;
|
|
bool dual_fund;
|
|
u8 *error;
|
|
struct peer_fd *peer_fd = new_peer_fd(tmpctx, fd);
|
|
|
|
if (!fromwire_connectd_peer_active(msg, msg, &id, &msgtype, &channel_id))
|
|
fatal("Connectd gave bad CONNECTD_PEER_ACTIVE message %s",
|
|
tal_hex(msg, msg));
|
|
|
|
peer = peer_by_id(ld, &id);
|
|
if (!peer) {
|
|
/* This race is possible, but I want to see it in CI. */
|
|
log_broken(ld->log, "Unknown active peer %s",
|
|
type_to_string(tmpctx, struct node_id, &id));
|
|
return;
|
|
}
|
|
|
|
/* Do we know what channel they're talking about? */
|
|
channel = find_channel_by_id(peer, &channel_id);
|
|
if (channel) {
|
|
/* If we have a canned error for this channel, send it now */
|
|
if (channel->error) {
|
|
error = channel->error;
|
|
goto send_error;
|
|
}
|
|
|
|
switch (channel->state) {
|
|
case ONCHAIN:
|
|
case FUNDING_SPEND_SEEN:
|
|
case CLOSINGD_COMPLETE:
|
|
goto channel_is_closed;
|
|
case CLOSED:
|
|
/* Channel should not have been loaded */
|
|
abort();
|
|
case AWAITING_UNILATERAL: {
|
|
/* channel->error is not saved in db, so this can
|
|
* happen if we restart. */
|
|
error = towire_errorfmt(tmpctx, &channel->cid,
|
|
"Awaiting unilateral close");
|
|
goto send_error;
|
|
}
|
|
case DUALOPEND_OPEN_INIT:
|
|
/* We asked for this, to open? */
|
|
if (!msgtype
|
|
&& channel->open_attempt
|
|
&& channel->open_attempt->open_msg) {
|
|
if (peer_start_dualopend(peer, peer_fd, channel))
|
|
subd_send_msg(channel->owner, channel->open_attempt->open_msg);
|
|
goto subd_setup_done;
|
|
}
|
|
/* Fall through. */
|
|
case DUALOPEND_AWAITING_LOCKIN:
|
|
assert(!channel->owner);
|
|
peer_restart_dualopend(peer, peer_fd, channel);
|
|
goto subd_setup_done;
|
|
case CHANNELD_AWAITING_LOCKIN:
|
|
case CHANNELD_NORMAL:
|
|
case CHANNELD_SHUTTING_DOWN:
|
|
case CLOSINGD_SIGEXCHANGE:
|
|
assert(!channel->owner);
|
|
peer_start_channeld(channel,
|
|
peer_fd,
|
|
NULL, true,
|
|
NULL);
|
|
goto subd_setup_done;
|
|
}
|
|
abort();
|
|
}
|
|
|
|
dual_fund = feature_negotiated(ld->our_features,
|
|
peer->their_features,
|
|
OPT_DUAL_FUND);
|
|
|
|
/* Did we ask for this? */
|
|
if (!msgtype) {
|
|
/* If it was dual_fund, it will have peer_unsaved_channel above */
|
|
if (dual_fund) {
|
|
log_broken(ld->log, "Unsolicited active df peer %s?",
|
|
type_to_string(tmpctx, struct node_id,
|
|
&peer->id));
|
|
} else {
|
|
const struct uncommitted_channel *uc
|
|
= peer->uncommitted_channel;
|
|
|
|
if (!uc->open_daemon
|
|
&& uc->fc
|
|
&& uc->fc->open_msg) {
|
|
if (peer_start_openingd(peer, peer_fd)) {
|
|
subd_send_msg(uc->open_daemon,
|
|
uc->fc->open_msg);
|
|
}
|
|
} else {
|
|
log_broken(ld->log, "Unsolicited active peer %s?",
|
|
type_to_string(tmpctx, struct node_id,
|
|
&peer->id));
|
|
}
|
|
}
|
|
return;
|
|
}
|
|
|
|
/* OK, it's an unknown channel. Create a new one if they're trying. */
|
|
switch (*msgtype) {
|
|
case WIRE_OPEN_CHANNEL:
|
|
if (dual_fund) {
|
|
error = towire_errorfmt(tmpctx, &channel_id,
|
|
"OPT_DUAL_FUND: cannot use open_channel");
|
|
goto send_error;
|
|
}
|
|
if (peer->uncommitted_channel) {
|
|
error = towire_errorfmt(tmpctx, &channel_id,
|
|
"Multiple simulteneous opens not supported");
|
|
goto send_error;
|
|
}
|
|
peer->uncommitted_channel = new_uncommitted_channel(peer);
|
|
peer_start_openingd(peer, peer_fd);
|
|
break;
|
|
case WIRE_OPEN_CHANNEL2:
|
|
if (!dual_fund) {
|
|
error = towire_errorfmt(tmpctx, &channel_id,
|
|
"Didn't negotiate OPT_DUAL_FUND: cannot use open_channel2");
|
|
goto send_error;
|
|
}
|
|
channel = new_unsaved_channel(peer,
|
|
peer->ld->config.fee_base,
|
|
peer->ld->config.fee_per_satoshi);
|
|
channel->cid = channel_id;
|
|
peer_start_dualopend(peer, peer_fd, channel);
|
|
break;
|
|
default:
|
|
log_peer_unusual(ld->log, &peer->id,
|
|
"Unknown channel %s for %s",
|
|
type_to_string(tmpctx, struct channel_id,
|
|
&channel_id),
|
|
peer_wire_name(*msgtype));
|
|
error = towire_errorfmt(tmpctx, &channel_id,
|
|
"Unknown channel for %s", peer_wire_name(*msgtype));
|
|
goto send_error;
|
|
break;
|
|
}
|
|
return;
|
|
|
|
channel_is_closed:
|
|
if (msgtype && *msgtype == WIRE_CHANNEL_REESTABLISH) {
|
|
log_debug(channel->log,
|
|
"Reestablish on %s channel: using channeld to reply",
|
|
channel_state_name(channel));
|
|
peer_start_channeld(channel, peer_fd, NULL, true, true);
|
|
return;
|
|
}
|
|
|
|
/* Retransmit error if we have one. Otherwise generic error. */
|
|
error = channel->error;
|
|
if (!error)
|
|
error = towire_errorfmt(tmpctx, &channel_id,
|
|
"channel in state %s",
|
|
channel_state_name(channel));
|
|
|
|
send_error:
|
|
log_peer_debug(ld->log, &peer->id, "Telling connectd to send error %s",
|
|
tal_hex(tmpctx, error));
|
|
/* Get connectd to send error and close. */
|
|
subd_send_msg(ld->connectd,
|
|
take(towire_connectd_peer_final_msg(NULL, &peer->id,
|
|
error)));
|
|
return;
|
|
|
|
subd_setup_done:
|
|
/* We deferred connect_succeeded to here, so subd would be ready once
|
|
* `connect` returns. */
|
|
if (!peer_subds_pending(peer))
|
|
connect_succeeded(ld, peer, peer->connected_incoming, &peer->addr);
|
|
}
|
|
|
|
struct disconnect_command {
|
|
struct list_node list;
|
|
/* Command structure. This is the parent of the close command. */
|
|
struct command *cmd;
|
|
/* node being disconnected. */
|
|
struct node_id id;
|
|
};
|
|
|
|
static void destroy_disconnect_command(struct disconnect_command *dc)
|
|
{
|
|
list_del(&dc->list);
|
|
}
|
|
|
|
void peer_disconnect_done(struct lightningd *ld, const u8 *msg)
|
|
{
|
|
struct node_id id;
|
|
struct disconnect_command *i, *next;
|
|
struct peer *p;
|
|
|
|
if (!fromwire_connectd_peer_disconnect_done(msg, &id))
|
|
fatal("Connectd gave bad PEER_DISCONNECT_DONE message %s",
|
|
tal_hex(msg, msg));
|
|
|
|
/* If we still have peer, it's disconnected now */
|
|
p = peer_by_id(ld, &id);
|
|
if (p) {
|
|
log_peer_debug(ld->log, &id, "peer_disconnect_done");
|
|
p->is_connected = false;
|
|
/* If we only cared about peer because of connectd, free it. */
|
|
if (list_empty(&p->channels) && !p->uncommitted_channel) {
|
|
tal_free(p);
|
|
}
|
|
}
|
|
|
|
/* Fire off plugin notifications */
|
|
notify_disconnect(ld, &id);
|
|
|
|
/* Wake any disconnect commands (removes self from list) */
|
|
list_for_each_safe(&ld->disconnect_commands, i, next, list) {
|
|
if (!node_id_eq(&i->id, &id))
|
|
continue;
|
|
|
|
was_pending(command_success(i->cmd,
|
|
json_stream_success(i->cmd)));
|
|
}
|
|
}
|
|
|
|
static bool check_funding_details(const struct bitcoin_tx *tx,
|
|
const u8 *wscript,
|
|
struct amount_sat funding,
|
|
u32 funding_outnum)
|
|
{
|
|
struct amount_asset asset =
|
|
bitcoin_tx_output_get_amount(tx, funding_outnum);
|
|
|
|
if (!amount_asset_is_main(&asset))
|
|
return false;
|
|
|
|
if (funding_outnum >= tx->wtx->num_outputs)
|
|
return false;
|
|
|
|
if (!amount_sat_eq(amount_asset_to_sat(&asset), funding))
|
|
return false;
|
|
|
|
return scripteq(scriptpubkey_p2wsh(tmpctx, wscript),
|
|
bitcoin_tx_output_get_script(tmpctx, tx,
|
|
funding_outnum));
|
|
}
|
|
|
|
|
|
/* FIXME: Unify our watch code so we get notified by txout, instead, like
|
|
* the wallet code does. */
|
|
static bool check_funding_tx(const struct bitcoin_tx *tx,
|
|
const struct channel *channel)
|
|
{
|
|
struct channel_inflight *inflight;
|
|
const u8 *wscript;
|
|
wscript = bitcoin_redeem_2of2(tmpctx,
|
|
&channel->local_funding_pubkey,
|
|
&channel->channel_info.remote_fundingkey);
|
|
|
|
/* Since we've enabled "RBF" for funding transactions,
|
|
* it's possible that it's one of "inflights".
|
|
* Worth noting that this check was added to prevent
|
|
* a peer from sending us a 'bogus' transaction id (that didn't
|
|
* actually contain the funding output). As of v2 (where
|
|
* RBF is introduced), this isn't a problem so much as
|
|
* both sides have full access to the funding transaction */
|
|
if (check_funding_details(tx, wscript, channel->funding_sats,
|
|
channel->funding.n))
|
|
return true;
|
|
|
|
list_for_each(&channel->inflights, inflight, list) {
|
|
if (check_funding_details(tx, wscript,
|
|
inflight->funding->total_funds,
|
|
inflight->funding->outpoint.n))
|
|
return true;
|
|
}
|
|
return false;
|
|
}
|
|
|
|
static void update_channel_from_inflight(struct lightningd *ld,
|
|
struct channel *channel,
|
|
const struct channel_inflight *inflight)
|
|
{
|
|
struct wally_psbt *psbt_copy;
|
|
|
|
channel->funding = inflight->funding->outpoint;
|
|
channel->funding_sats = inflight->funding->total_funds;
|
|
channel->our_funds = inflight->funding->our_funds;
|
|
|
|
/* Lease infos ! */
|
|
channel->lease_expiry = inflight->lease_expiry;
|
|
channel->push = inflight->lease_fee;
|
|
tal_free(channel->lease_commit_sig);
|
|
channel->lease_commit_sig
|
|
= tal_dup_or_null(channel, secp256k1_ecdsa_signature, inflight->lease_commit_sig);
|
|
channel->lease_chan_max_msat = inflight->lease_chan_max_msat;
|
|
channel->lease_chan_max_ppt = inflight->lease_chan_max_ppt;
|
|
|
|
tal_free(channel->blockheight_states);
|
|
channel->blockheight_states = new_height_states(channel,
|
|
channel->opener,
|
|
&inflight->lease_blockheight_start);
|
|
|
|
/* Make a 'clone' of this tx */
|
|
psbt_copy = clone_psbt(channel, inflight->last_tx->psbt);
|
|
channel_set_last_tx(channel,
|
|
bitcoin_tx_with_psbt(channel, psbt_copy),
|
|
&inflight->last_sig,
|
|
TX_CHANNEL_UNILATERAL);
|
|
|
|
/* Update the reserve */
|
|
channel_update_reserve(channel,
|
|
&channel->channel_info.their_config,
|
|
inflight->funding->total_funds);
|
|
|
|
wallet_channel_save(ld->wallet, channel);
|
|
}
|
|
|
|
static enum watch_result funding_depth_cb(struct lightningd *ld,
|
|
struct channel *channel,
|
|
const struct bitcoin_txid *txid,
|
|
const struct bitcoin_tx *tx,
|
|
unsigned int depth)
|
|
{
|
|
const char *txidstr;
|
|
struct short_channel_id scid;
|
|
|
|
/* Sanity check */
|
|
if (!check_funding_tx(tx, channel)) {
|
|
channel_internal_error(channel, "Bad tx %s: %s",
|
|
type_to_string(tmpctx,
|
|
struct bitcoin_txid, txid),
|
|
type_to_string(tmpctx,
|
|
struct bitcoin_tx, tx));
|
|
return DELETE_WATCH;
|
|
}
|
|
|
|
txidstr = type_to_string(tmpctx, struct bitcoin_txid, txid);
|
|
log_debug(channel->log, "Funding tx %s depth %u of %u",
|
|
txidstr, depth, channel->minimum_depth);
|
|
tal_free(txidstr);
|
|
|
|
bool min_depth_reached = depth >= channel->minimum_depth;
|
|
|
|
/* Reorg can change scid, so always update/save scid when possible (depth=0
|
|
* means the stale block with our funding tx was removed) */
|
|
if ((min_depth_reached && !channel->scid) || (depth && channel->scid)) {
|
|
struct txlocator *loc;
|
|
struct channel_inflight *inf;
|
|
|
|
/* Update the channel's info to the correct tx, if needed to
|
|
* It's possible an 'inflight' has reached depth */
|
|
if (!list_empty(&channel->inflights)) {
|
|
inf = channel_inflight_find(channel, txid);
|
|
if (!inf) {
|
|
channel_fail_permanent(channel, REASON_LOCAL,
|
|
"Txid %s for channel"
|
|
" not found in inflights. (peer %s)",
|
|
type_to_string(tmpctx,
|
|
struct bitcoin_txid,
|
|
txid),
|
|
type_to_string(tmpctx,
|
|
struct node_id,
|
|
&channel->peer->id));
|
|
return DELETE_WATCH;
|
|
}
|
|
update_channel_from_inflight(ld, channel, inf);
|
|
}
|
|
|
|
wallet_annotate_txout(ld->wallet, &channel->funding,
|
|
TX_CHANNEL_FUNDING, channel->dbid);
|
|
loc = wallet_transaction_locate(tmpctx, ld->wallet, txid);
|
|
if (!mk_short_channel_id(&scid,
|
|
loc->blkheight, loc->index,
|
|
channel->funding.n)) {
|
|
channel_fail_permanent(channel,
|
|
REASON_LOCAL,
|
|
"Invalid funding scid %u:%u:%u",
|
|
loc->blkheight, loc->index,
|
|
channel->funding.n);
|
|
return DELETE_WATCH;
|
|
}
|
|
|
|
/* If we restart, we could already have peer->scid from database */
|
|
if (!channel->scid) {
|
|
channel->scid = tal(channel, struct short_channel_id);
|
|
*channel->scid = scid;
|
|
wallet_channel_save(ld->wallet, channel);
|
|
|
|
} else if (!short_channel_id_eq(channel->scid, &scid)) {
|
|
/* This normally restarts channeld, initialized with updated scid
|
|
* and also adds it (at least our halve_chan) to rtable. */
|
|
channel_fail_reconnect(channel,
|
|
"short_channel_id changed to %s (was %s)",
|
|
short_channel_id_to_str(tmpctx, &scid),
|
|
short_channel_id_to_str(tmpctx, channel->scid));
|
|
|
|
*channel->scid = scid;
|
|
wallet_channel_save(ld->wallet, channel);
|
|
return KEEP_WATCHING;
|
|
}
|
|
}
|
|
|
|
/* Try to tell subdaemon */
|
|
if (!channel_tell_depth(ld, channel, txid, depth))
|
|
return KEEP_WATCHING;
|
|
|
|
if (!min_depth_reached)
|
|
return KEEP_WATCHING;
|
|
|
|
/* We keep telling it depth/scid until we get to announce depth. */
|
|
if (depth < ANNOUNCE_MIN_DEPTH)
|
|
return KEEP_WATCHING;
|
|
|
|
return DELETE_WATCH;
|
|
}
|
|
|
|
static enum watch_result funding_spent(struct channel *channel,
|
|
const struct bitcoin_tx *tx,
|
|
size_t inputnum UNUSED,
|
|
const struct block *block)
|
|
{
|
|
struct bitcoin_txid txid;
|
|
bitcoin_txid(tx, &txid);
|
|
|
|
wallet_channeltxs_add(channel->peer->ld->wallet, channel,
|
|
WIRE_ONCHAIND_INIT, &txid, 0, block->height);
|
|
return onchaind_funding_spent(channel, tx, block->height);
|
|
}
|
|
|
|
void channel_watch_wrong_funding(struct lightningd *ld, struct channel *channel)
|
|
{
|
|
/* Watch the "wrong" funding too, in case we spend it. */
|
|
if (channel->shutdown_wrong_funding) {
|
|
/* FIXME: Remove arg from cb? */
|
|
watch_txo(channel, ld->topology, channel,
|
|
channel->shutdown_wrong_funding,
|
|
funding_spent);
|
|
}
|
|
}
|
|
|
|
void channel_watch_funding(struct lightningd *ld, struct channel *channel)
|
|
{
|
|
/* FIXME: Remove arg from cb? */
|
|
watch_txid(channel, ld->topology, channel,
|
|
&channel->funding.txid, funding_depth_cb);
|
|
watch_txo(channel, ld->topology, channel,
|
|
&channel->funding,
|
|
funding_spent);
|
|
channel_watch_wrong_funding(ld, channel);
|
|
}
|
|
|
|
static void channel_watch_inflight(struct lightningd *ld,
|
|
struct channel *channel,
|
|
struct channel_inflight *inflight)
|
|
{
|
|
/* FIXME: Remove arg from cb? */
|
|
watch_txid(channel, ld->topology, channel,
|
|
&inflight->funding->outpoint.txid, funding_depth_cb);
|
|
watch_txo(channel, ld->topology, channel,
|
|
&inflight->funding->outpoint,
|
|
funding_spent);
|
|
}
|
|
|
|
static void json_add_peer(struct lightningd *ld,
|
|
struct json_stream *response,
|
|
struct peer *p,
|
|
const enum log_level *ll)
|
|
{
|
|
struct channel *channel;
|
|
|
|
json_object_start(response, NULL);
|
|
json_add_node_id(response, "id", &p->id);
|
|
|
|
json_add_bool(response, "connected", p->is_connected);
|
|
|
|
/* If it's not connected, features are unreliable: we don't
|
|
* store them in the database, and they would only reflect
|
|
* their features *last* time they connected. */
|
|
if (p->is_connected) {
|
|
json_array_start(response, "netaddr");
|
|
json_add_string(response, NULL,
|
|
type_to_string(tmpctx,
|
|
struct wireaddr_internal,
|
|
&p->addr));
|
|
json_array_end(response);
|
|
json_add_hex_talarr(response, "features", p->their_features);
|
|
}
|
|
|
|
json_array_start(response, "channels");
|
|
json_add_uncommitted_channel(response, p->uncommitted_channel);
|
|
|
|
list_for_each(&p->channels, channel, list) {
|
|
if (channel_unsaved(channel))
|
|
json_add_unsaved_channel(response, channel);
|
|
else
|
|
json_add_channel(ld, response, NULL, channel);
|
|
}
|
|
json_array_end(response);
|
|
|
|
if (ll)
|
|
json_add_log(response, ld->log_book, &p->id, *ll);
|
|
json_object_end(response);
|
|
}
|
|
|
|
static struct command_result *json_listpeers(struct command *cmd,
|
|
const char *buffer,
|
|
const jsmntok_t *obj UNNEEDED,
|
|
const jsmntok_t *params)
|
|
{
|
|
enum log_level *ll;
|
|
struct node_id *specific_id;
|
|
struct peer *peer;
|
|
struct json_stream *response;
|
|
|
|
if (!param(cmd, buffer, params,
|
|
p_opt("id", param_node_id, &specific_id),
|
|
p_opt("level", param_loglevel, &ll),
|
|
NULL))
|
|
return command_param_failed();
|
|
|
|
response = json_stream_success(cmd);
|
|
json_array_start(response, "peers");
|
|
if (specific_id) {
|
|
peer = peer_by_id(cmd->ld, specific_id);
|
|
if (peer)
|
|
json_add_peer(cmd->ld, response, peer, ll);
|
|
} else {
|
|
list_for_each(&cmd->ld->peers, peer, list)
|
|
json_add_peer(cmd->ld, response, peer, ll);
|
|
}
|
|
json_array_end(response);
|
|
|
|
return command_success(cmd, response);
|
|
}
|
|
|
|
static const struct json_command listpeers_command = {
|
|
"listpeers",
|
|
"network",
|
|
json_listpeers,
|
|
"Show current peers, if {level} is set, include logs for {id}"
|
|
};
|
|
/* Comment added to satisfice AUTODATA */
|
|
AUTODATA(json_command, &listpeers_command);
|
|
|
|
struct command_result *
|
|
command_find_channel(struct command *cmd,
|
|
const char *buffer, const jsmntok_t *tok,
|
|
struct channel **channel)
|
|
{
|
|
struct lightningd *ld = cmd->ld;
|
|
struct channel_id cid;
|
|
struct short_channel_id scid;
|
|
struct peer *peer;
|
|
|
|
if (json_tok_channel_id(buffer, tok, &cid)) {
|
|
list_for_each(&ld->peers, peer, list) {
|
|
list_for_each(&peer->channels, (*channel), list) {
|
|
if (!channel_active(*channel))
|
|
continue;
|
|
if (channel_id_eq(&(*channel)->cid, &cid))
|
|
return NULL;
|
|
}
|
|
}
|
|
return command_fail(cmd, JSONRPC2_INVALID_PARAMS,
|
|
"Channel ID not found: '%.*s'",
|
|
tok->end - tok->start,
|
|
buffer + tok->start);
|
|
} else if (json_to_short_channel_id(buffer, tok, &scid)) {
|
|
*channel = any_channel_by_scid(ld, &scid);
|
|
if (!*channel)
|
|
return command_fail(cmd, JSONRPC2_INVALID_PARAMS,
|
|
"Short channel ID not found: '%.*s'",
|
|
tok->end - tok->start,
|
|
buffer + tok->start);
|
|
if (!channel_active(*channel))
|
|
return command_fail(cmd, JSONRPC2_INVALID_PARAMS,
|
|
"Short channel ID not active: '%.*s'",
|
|
tok->end - tok->start,
|
|
buffer + tok->start);
|
|
return NULL;
|
|
} else {
|
|
return command_fail_badparam(cmd, "id", buffer, tok,
|
|
"should be a channel ID or short channel ID");
|
|
}
|
|
}
|
|
|
|
static void setup_peer(struct peer *peer, u32 delay)
|
|
{
|
|
struct channel *channel;
|
|
struct channel_inflight *inflight;
|
|
struct lightningd *ld = peer->ld;
|
|
bool connect = false;
|
|
|
|
list_for_each(&peer->channels, channel, list) {
|
|
if (channel_unsaved(channel))
|
|
continue;
|
|
/* Watching lockin may be unnecessary, but it's harmless. */
|
|
channel_watch_funding(ld, channel);
|
|
|
|
/* Also watch any inflight txs */
|
|
list_for_each(&channel->inflights, inflight, list) {
|
|
/* Don't double watch the txid that's also in
|
|
* channel->funding_txid */
|
|
if (bitcoin_txid_eq(&channel->funding.txid,
|
|
&inflight->funding->outpoint.txid))
|
|
continue;
|
|
|
|
channel_watch_inflight(ld, channel, inflight);
|
|
}
|
|
if (channel_active(channel))
|
|
connect = true;
|
|
}
|
|
|
|
/* Make sure connectd knows to try reconnecting. */
|
|
if (connect)
|
|
try_reconnect(peer, peer, delay, &peer->addr);
|
|
|
|
}
|
|
|
|
void setup_peers(struct lightningd *ld)
|
|
{
|
|
struct peer *p;
|
|
/* Avoid thundering herd: after first five, delay by 1 second. */
|
|
int delay = -5;
|
|
|
|
list_for_each(&ld->peers, p, list) {
|
|
setup_peer(p, delay > 0 ? delay : 0);
|
|
delay++;
|
|
}
|
|
}
|
|
|
|
/* Pull peers, channels and HTLCs from db, and wire them up. */
|
|
struct htlc_in_map *load_channels_from_wallet(struct lightningd *ld)
|
|
{
|
|
struct peer *peer;
|
|
struct htlc_in_map *unconnected_htlcs_in = tal(ld, struct htlc_in_map);
|
|
|
|
/* Load channels from database */
|
|
if (!wallet_init_channels(ld->wallet))
|
|
fatal("Could not load channels from the database");
|
|
|
|
/* First we load the incoming htlcs */
|
|
list_for_each(&ld->peers, peer, list) {
|
|
struct channel *channel;
|
|
|
|
list_for_each(&peer->channels, channel, list) {
|
|
if (!wallet_htlcs_load_in_for_channel(ld->wallet,
|
|
channel,
|
|
&ld->htlcs_in)) {
|
|
fatal("could not load htlcs for channel");
|
|
}
|
|
}
|
|
}
|
|
|
|
/* Make a copy of the htlc_map: entries removed as they're matched */
|
|
htlc_in_map_copy(unconnected_htlcs_in, &ld->htlcs_in);
|
|
|
|
/* Now we load the outgoing HTLCs, so we can connect them. */
|
|
list_for_each(&ld->peers, peer, list) {
|
|
struct channel *channel;
|
|
|
|
list_for_each(&peer->channels, channel, list) {
|
|
if (!wallet_htlcs_load_out_for_channel(ld->wallet,
|
|
channel,
|
|
&ld->htlcs_out,
|
|
unconnected_htlcs_in)) {
|
|
fatal("could not load outgoing htlcs for channel");
|
|
}
|
|
}
|
|
}
|
|
|
|
#ifdef COMPAT_V061
|
|
fixup_htlcs_out(ld);
|
|
#endif /* COMPAT_V061 */
|
|
|
|
return unconnected_htlcs_in;
|
|
}
|
|
|
|
static struct command_result *json_disconnect(struct command *cmd,
|
|
const char *buffer,
|
|
const jsmntok_t *obj UNNEEDED,
|
|
const jsmntok_t *params)
|
|
{
|
|
struct node_id *id;
|
|
struct disconnect_command *dc;
|
|
struct peer *peer;
|
|
struct channel *channel, **channels;
|
|
bool *force;
|
|
bool disconnected = false;
|
|
|
|
if (!param(cmd, buffer, params,
|
|
p_req("id", param_node_id, &id),
|
|
p_opt_def("force", param_bool, &force, false),
|
|
NULL))
|
|
return command_param_failed();
|
|
|
|
peer = peer_by_id(cmd->ld, id);
|
|
if (!peer) {
|
|
return command_fail(cmd, LIGHTNINGD, "Unknown peer");
|
|
}
|
|
if (!peer->is_connected) {
|
|
return command_fail(cmd, LIGHTNINGD, "Peer not connected");
|
|
}
|
|
|
|
channel = peer_any_active_channel(peer, NULL);
|
|
if (channel && !*force) {
|
|
return command_fail(cmd, LIGHTNINGD,
|
|
"Peer has (at least one) channel in state %s",
|
|
channel_state_name(channel));
|
|
}
|
|
|
|
/* Careful here! Disconnecting can free peer! */
|
|
channels = tal_arr(cmd, struct channel *, 0);
|
|
list_for_each(&peer->channels, channel, list) {
|
|
if (!channel->owner)
|
|
continue;
|
|
if (!channel->owner->talks_to_peer)
|
|
continue;
|
|
|
|
switch (channel->state) {
|
|
case DUALOPEND_OPEN_INIT:
|
|
case CHANNELD_AWAITING_LOCKIN:
|
|
case CHANNELD_NORMAL:
|
|
case CHANNELD_SHUTTING_DOWN:
|
|
case DUALOPEND_AWAITING_LOCKIN:
|
|
case CLOSINGD_SIGEXCHANGE:
|
|
tal_arr_expand(&channels, channel);
|
|
continue;
|
|
case CLOSINGD_COMPLETE:
|
|
case AWAITING_UNILATERAL:
|
|
case FUNDING_SPEND_SEEN:
|
|
case ONCHAIN:
|
|
case CLOSED:
|
|
/* We don't expect these to have owners who connect! */
|
|
log_broken(channel->log,
|
|
"Don't expect owner %s in state %s",
|
|
channel->owner->name,
|
|
channel_state_name(channel));
|
|
continue;
|
|
}
|
|
abort();
|
|
}
|
|
|
|
/* This can free peer too! */
|
|
if (peer->uncommitted_channel) {
|
|
kill_uncommitted_channel(peer->uncommitted_channel,
|
|
"disconnect command");
|
|
disconnected = true;
|
|
}
|
|
|
|
for (size_t i = 0; i < tal_count(channels); i++) {
|
|
if (channel_unsaved(channels[i]))
|
|
channel_unsaved_close_conn(channels[i],
|
|
"disconnect command");
|
|
else
|
|
channel_fail_reconnect(channels[i],
|
|
"disconnect command");
|
|
disconnected = true;
|
|
}
|
|
|
|
/* It's just sitting in connectd? */
|
|
if (!disconnected)
|
|
maybe_disconnect_peer(cmd->ld, peer);
|
|
|
|
/* Connectd tells us when it's finally disconnected */
|
|
dc = tal(cmd, struct disconnect_command);
|
|
dc->cmd = cmd;
|
|
dc->id = *id;
|
|
list_add_tail(&cmd->ld->disconnect_commands, &dc->list);
|
|
tal_add_destructor(dc, destroy_disconnect_command);
|
|
|
|
return command_still_pending(cmd);
|
|
}
|
|
|
|
static const struct json_command disconnect_command = {
|
|
"disconnect",
|
|
"network",
|
|
json_disconnect,
|
|
"Disconnect from {id} that has previously been connected to using connect; with {force} set, even if it has a current channel"
|
|
};
|
|
AUTODATA(json_command, &disconnect_command);
|
|
|
|
static struct command_result *json_getinfo(struct command *cmd,
|
|
const char *buffer,
|
|
const jsmntok_t *obj UNNEEDED,
|
|
const jsmntok_t *params)
|
|
{
|
|
struct json_stream *response;
|
|
struct peer *peer;
|
|
struct channel *channel;
|
|
unsigned int pending_channels = 0, active_channels = 0,
|
|
inactive_channels = 0, num_peers = 0;
|
|
|
|
if (!param(cmd, buffer, params, NULL))
|
|
return command_param_failed();
|
|
|
|
response = json_stream_success(cmd);
|
|
json_add_node_id(response, "id", &cmd->ld->id);
|
|
json_add_string(response, "alias", (const char *)cmd->ld->alias);
|
|
json_add_hex_talarr(response, "color", cmd->ld->rgb);
|
|
|
|
/* Add some peer and channel stats */
|
|
list_for_each(&cmd->ld->peers, peer, list) {
|
|
num_peers++;
|
|
|
|
list_for_each(&peer->channels, channel, list) {
|
|
if (channel->state == CHANNELD_AWAITING_LOCKIN
|
|
|| channel->state == DUALOPEND_AWAITING_LOCKIN
|
|
|| channel->state == DUALOPEND_OPEN_INIT) {
|
|
pending_channels++;
|
|
} else if (channel_active(channel)) {
|
|
active_channels++;
|
|
} else {
|
|
inactive_channels++;
|
|
}
|
|
}
|
|
}
|
|
json_add_num(response, "num_peers", num_peers);
|
|
json_add_num(response, "num_pending_channels", pending_channels);
|
|
json_add_num(response, "num_active_channels", active_channels);
|
|
json_add_num(response, "num_inactive_channels", inactive_channels);
|
|
|
|
/* Add network info */
|
|
if (cmd->ld->listen) {
|
|
/* These are the addresses we're announcing */
|
|
json_array_start(response, "address");
|
|
for (size_t i = 0; i < tal_count(cmd->ld->announceable); i++)
|
|
json_add_address(response, NULL, cmd->ld->announceable+i);
|
|
json_array_end(response);
|
|
|
|
/* This is what we're actually bound to. */
|
|
json_array_start(response, "binding");
|
|
for (size_t i = 0; i < tal_count(cmd->ld->binding); i++)
|
|
json_add_address_internal(response, NULL,
|
|
cmd->ld->binding+i);
|
|
json_array_end(response);
|
|
}
|
|
json_add_string(response, "version", version());
|
|
json_add_num(response, "blockheight", cmd->ld->blockheight);
|
|
json_add_string(response, "network", chainparams->network_name);
|
|
json_add_amount_msat_compat(response,
|
|
wallet_total_forward_fees(cmd->ld->wallet),
|
|
"msatoshi_fees_collected",
|
|
"fees_collected_msat");
|
|
json_add_string(response, "lightning-dir", cmd->ld->config_netdir);
|
|
|
|
if (!cmd->ld->topology->bitcoind->synced)
|
|
json_add_string(response, "warning_bitcoind_sync",
|
|
"Bitcoind is not up-to-date with network.");
|
|
else if (!topology_synced(cmd->ld->topology))
|
|
json_add_string(response, "warning_lightningd_sync",
|
|
"Still loading latest blocks from bitcoind.");
|
|
|
|
u8 **bits = cmd->ld->our_features->bits;
|
|
json_object_start(response, "our_features");
|
|
json_add_hex_talarr(response, "init",
|
|
featurebits_or(cmd, bits[INIT_FEATURE], bits[GLOBAL_INIT_FEATURE]));
|
|
json_add_hex_talarr(response, "node", bits[NODE_ANNOUNCE_FEATURE]);
|
|
json_add_hex_talarr(response, "channel", bits[CHANNEL_FEATURE]);
|
|
json_add_hex_talarr(response, "invoice", bits[BOLT11_FEATURE]);
|
|
json_object_end(response);
|
|
|
|
return command_success(cmd, response);
|
|
}
|
|
|
|
static const struct json_command getinfo_command = {
|
|
"getinfo",
|
|
"utility",
|
|
json_getinfo,
|
|
"Show information about this node"
|
|
};
|
|
AUTODATA(json_command, &getinfo_command);
|
|
|
|
/* Wait for at least a specific blockheight, then return, or time out. */
|
|
struct waitblockheight_waiter {
|
|
/* struct lightningd::waitblockheight_commands. */
|
|
struct list_node list;
|
|
/* Command structure. This is the parent of the close command. */
|
|
struct command *cmd;
|
|
/* The block height being waited for. */
|
|
u32 block_height;
|
|
/* Whether we have been removed from the list. */
|
|
bool removed;
|
|
};
|
|
/* Completes a pending waitblockheight. */
|
|
static struct command_result *
|
|
waitblockheight_complete(struct command *cmd,
|
|
u32 block_height)
|
|
{
|
|
struct json_stream *response;
|
|
|
|
response = json_stream_success(cmd);
|
|
json_add_num(response, "blockheight", block_height);
|
|
return command_success(cmd, response);
|
|
}
|
|
/* Called when command is destroyed without being resolved. */
|
|
static void
|
|
destroy_waitblockheight_waiter(struct waitblockheight_waiter *w)
|
|
{
|
|
if (!w->removed)
|
|
list_del(&w->list);
|
|
}
|
|
/* Called on timeout. */
|
|
static void
|
|
timeout_waitblockheight_waiter(struct waitblockheight_waiter *w)
|
|
{
|
|
list_del(&w->list);
|
|
w->removed = true;
|
|
tal_steal(tmpctx, w);
|
|
was_pending(command_fail(w->cmd, WAIT_TIMEOUT,
|
|
"Timed out."));
|
|
}
|
|
/* Called by lightningd at each new block. */
|
|
void waitblockheight_notify_new_block(struct lightningd *ld,
|
|
u32 block_height)
|
|
{
|
|
struct waitblockheight_waiter *w, *n;
|
|
char *to_delete = tal(NULL, char);
|
|
|
|
/* Use safe since we could resolve commands and thus
|
|
* trigger removal of list elements.
|
|
*/
|
|
list_for_each_safe(&ld->waitblockheight_commands, w, n, list) {
|
|
/* Skip commands that have not been reached yet. */
|
|
if (w->block_height > block_height)
|
|
continue;
|
|
|
|
list_del(&w->list);
|
|
w->removed = true;
|
|
tal_steal(to_delete, w);
|
|
was_pending(waitblockheight_complete(w->cmd,
|
|
block_height));
|
|
}
|
|
tal_free(to_delete);
|
|
}
|
|
static struct command_result *json_waitblockheight(struct command *cmd,
|
|
const char *buffer,
|
|
const jsmntok_t *obj,
|
|
const jsmntok_t *params)
|
|
{
|
|
unsigned int *target_block_height;
|
|
u32 block_height;
|
|
unsigned int *timeout;
|
|
struct waitblockheight_waiter *w;
|
|
|
|
if (!param(cmd, buffer, params,
|
|
p_req("blockheight", param_number, &target_block_height),
|
|
p_opt_def("timeout", param_number, &timeout, 60),
|
|
NULL))
|
|
return command_param_failed();
|
|
|
|
/* Check if already reached anyway. */
|
|
block_height = get_block_height(cmd->ld->topology);
|
|
if (*target_block_height <= block_height)
|
|
return waitblockheight_complete(cmd, block_height);
|
|
|
|
/* Create a new waitblockheight command. */
|
|
w = tal(cmd, struct waitblockheight_waiter);
|
|
tal_add_destructor(w, &destroy_waitblockheight_waiter);
|
|
list_add(&cmd->ld->waitblockheight_commands, &w->list);
|
|
w->cmd = cmd;
|
|
w->block_height = *target_block_height;
|
|
w->removed = false;
|
|
/* Install the timeout. */
|
|
(void) new_reltimer(cmd->ld->timers, w, time_from_sec(*timeout),
|
|
&timeout_waitblockheight_waiter, w);
|
|
|
|
return command_still_pending(cmd);
|
|
}
|
|
|
|
static const struct json_command waitblockheight_command = {
|
|
"waitblockheight",
|
|
"utility",
|
|
&json_waitblockheight,
|
|
"Wait for the blockchain to reach {blockheight}, up to "
|
|
"{timeout} seconds."
|
|
};
|
|
AUTODATA(json_command, &waitblockheight_command);
|
|
|
|
static struct command_result *param_channel_or_all(struct command *cmd,
|
|
const char *name,
|
|
const char *buffer,
|
|
const jsmntok_t *tok,
|
|
struct channel ***channels)
|
|
{
|
|
struct command_result *res;
|
|
struct peer *peer;
|
|
|
|
/* early return the easy case */
|
|
if (json_tok_streq(buffer, tok, "all")) {
|
|
*channels = NULL;
|
|
return NULL;
|
|
}
|
|
|
|
/* Find channels by peer_id */
|
|
peer = peer_from_json(cmd->ld, buffer, tok);
|
|
if (peer) {
|
|
struct channel *channel;
|
|
*channels = tal_arr(cmd, struct channel *, 0);
|
|
list_for_each(&peer->channels, channel, list) {
|
|
if (channel->state != CHANNELD_NORMAL
|
|
&& channel->state != CHANNELD_AWAITING_LOCKIN
|
|
&& channel->state != DUALOPEND_AWAITING_LOCKIN)
|
|
continue;
|
|
|
|
tal_arr_expand(channels, channel);
|
|
}
|
|
if (tal_count(*channels) == 0)
|
|
return command_fail(cmd, LIGHTNINGD,
|
|
"Could not find any active channels of peer with that id");
|
|
return NULL;
|
|
/* Find channel by id or scid */
|
|
} else {
|
|
struct channel *channel;
|
|
res = command_find_channel(cmd, buffer, tok, &channel);
|
|
if (res)
|
|
return res;
|
|
/* check channel is found and in valid state */
|
|
if (!channel)
|
|
return command_fail(cmd, LIGHTNINGD,
|
|
"Could not find channel with that id");
|
|
*channels = tal_arr(cmd, struct channel *, 1);
|
|
(*channels)[0] = channel;
|
|
return NULL;
|
|
}
|
|
}
|
|
|
|
/* Fee base is a u32, but it's convenient to let them specify it using
|
|
* msat etc. suffix. */
|
|
static struct command_result *param_msat_u32(struct command *cmd,
|
|
const char *name,
|
|
const char *buffer,
|
|
const jsmntok_t *tok,
|
|
u32 **num)
|
|
{
|
|
struct amount_msat *msat;
|
|
struct command_result *res;
|
|
|
|
/* Parse just like an msat. */
|
|
res = param_msat(cmd, name, buffer, tok, &msat);
|
|
if (res)
|
|
return res;
|
|
|
|
*num = tal(cmd, u32);
|
|
if (!amount_msat_to_u32(*msat, *num)) {
|
|
return command_fail_badparam(cmd, name, buffer, tok,
|
|
"exceeds u32 max");
|
|
}
|
|
|
|
return NULL;
|
|
}
|
|
|
|
static void set_channel_config(struct command *cmd, struct channel *channel,
|
|
u32 *base,
|
|
u32 *ppm,
|
|
struct amount_msat *htlc_min,
|
|
struct amount_msat *htlc_max,
|
|
u32 delaysecs,
|
|
struct json_stream *response,
|
|
bool add_details)
|
|
{
|
|
bool warn_cannot_set_min = false, warn_cannot_set_max = false;
|
|
|
|
/* We only need to defer values if we *increase* fees (or drop
|
|
* max, increase min); we always allow users to overpay fees. */
|
|
if ((base && *base > channel->feerate_base)
|
|
|| (ppm && *ppm > channel->feerate_ppm)
|
|
|| (htlc_min
|
|
&& amount_msat_greater(*htlc_min, channel->htlc_minimum_msat))
|
|
|| (htlc_max
|
|
&& amount_msat_less(*htlc_max, channel->htlc_maximum_msat))) {
|
|
channel->old_feerate_timeout
|
|
= timeabs_add(time_now(), time_from_sec(delaysecs));
|
|
channel->old_feerate_base = channel->feerate_base;
|
|
channel->old_feerate_ppm = channel->feerate_ppm;
|
|
channel->old_htlc_minimum_msat = channel->htlc_minimum_msat;
|
|
channel->old_htlc_maximum_msat = channel->htlc_maximum_msat;
|
|
}
|
|
|
|
/* set new values */
|
|
if (base)
|
|
channel->feerate_base = *base;
|
|
if (ppm)
|
|
channel->feerate_ppm = *ppm;
|
|
if (htlc_min) {
|
|
struct amount_msat actual_min;
|
|
|
|
/* We can't send something they'll refuse: check that here. */
|
|
actual_min = channel->channel_info.their_config.htlc_minimum;
|
|
if (amount_msat_less(*htlc_min, actual_min)) {
|
|
warn_cannot_set_min = true;
|
|
channel->htlc_minimum_msat = actual_min;
|
|
} else
|
|
channel->htlc_minimum_msat = *htlc_min;
|
|
}
|
|
if (htlc_max) {
|
|
struct amount_msat actual_max;
|
|
|
|
/* Can't set it greater than actual capacity. */
|
|
actual_max = htlc_max_possible_send(channel);
|
|
if (amount_msat_greater(*htlc_max, actual_max)) {
|
|
warn_cannot_set_max = true;
|
|
channel->htlc_maximum_msat = actual_max;
|
|
} else
|
|
channel->htlc_maximum_msat = *htlc_max;
|
|
}
|
|
|
|
/* tell channeld to make a send_channel_update */
|
|
if (channel->owner && streq(channel->owner->name, "channeld"))
|
|
subd_send_msg(channel->owner,
|
|
take(towire_channeld_config_channel(NULL, base, ppm,
|
|
htlc_min, htlc_max)));
|
|
|
|
/* save values to database */
|
|
wallet_channel_save(cmd->ld->wallet, channel);
|
|
|
|
/* write JSON response entry */
|
|
json_object_start(response, NULL);
|
|
json_add_node_id(response, "peer_id", &channel->peer->id);
|
|
json_add_string(response, "channel_id",
|
|
type_to_string(tmpctx, struct channel_id, &channel->cid));
|
|
if (channel->scid)
|
|
json_add_short_channel_id(response, "short_channel_id", channel->scid);
|
|
|
|
/* setchannel lists these explicitly */
|
|
if (add_details) {
|
|
json_add_amount_msat_only(response, "fee_base_msat",
|
|
amount_msat(channel->feerate_base));
|
|
json_add_u32(response, "fee_proportional_millionths",
|
|
channel->feerate_ppm);
|
|
json_add_amount_msat_only(response,
|
|
"minimum_htlc_out_msat",
|
|
channel->htlc_minimum_msat);
|
|
if (warn_cannot_set_min)
|
|
json_add_string(response, "warning_htlcmin_too_low",
|
|
"Set minimum_htlc_out_msat to minimum allowed by peer");
|
|
json_add_amount_msat_only(response,
|
|
"maximum_htlc_out_msat",
|
|
channel->htlc_maximum_msat);
|
|
if (warn_cannot_set_max)
|
|
json_add_string(response, "warning_htlcmax_too_high",
|
|
"Set maximum_htlc_out_msat to maximum possible in channel");
|
|
}
|
|
json_object_end(response);
|
|
}
|
|
|
|
static struct command_result *json_setchannelfee(struct command *cmd,
|
|
const char *buffer,
|
|
const jsmntok_t *obj UNNEEDED,
|
|
const jsmntok_t *params)
|
|
{
|
|
struct json_stream *response;
|
|
struct peer *peer;
|
|
struct channel **channels;
|
|
u32 *base, *ppm, *delaysecs;
|
|
|
|
/* Parse the JSON command */
|
|
if (!param(cmd, buffer, params,
|
|
p_req("id", param_channel_or_all, &channels),
|
|
p_opt_def("base", param_msat_u32,
|
|
&base, cmd->ld->config.fee_base),
|
|
p_opt_def("ppm", param_number, &ppm,
|
|
cmd->ld->config.fee_per_satoshi),
|
|
p_opt_def("enforcedelay", param_number, &delaysecs, 600),
|
|
NULL))
|
|
return command_param_failed();
|
|
|
|
/* Open JSON response object for later iteration */
|
|
response = json_stream_success(cmd);
|
|
json_add_num(response, "base", *base);
|
|
json_add_num(response, "ppm", *ppm);
|
|
json_array_start(response, "channels");
|
|
|
|
/* If the users requested 'all' channels we need to iterate */
|
|
if (channels == NULL) {
|
|
list_for_each(&cmd->ld->peers, peer, list) {
|
|
struct channel *channel;
|
|
list_for_each(&peer->channels, channel, list) {
|
|
if (channel->state != CHANNELD_NORMAL &&
|
|
channel->state != CHANNELD_AWAITING_LOCKIN &&
|
|
channel->state != DUALOPEND_AWAITING_LOCKIN)
|
|
continue;
|
|
set_channel_config(cmd, channel, base, ppm, NULL, NULL,
|
|
*delaysecs, response, false);
|
|
}
|
|
}
|
|
/* single peer should be updated */
|
|
} else {
|
|
for (size_t i = 0; i < tal_count(channels); i++) {
|
|
set_channel_config(cmd, channels[i], base, ppm, NULL, NULL,
|
|
*delaysecs, response, false);
|
|
}
|
|
}
|
|
|
|
/* Close and return response */
|
|
json_array_end(response);
|
|
return command_success(cmd, response);
|
|
}
|
|
|
|
static const struct json_command setchannelfee_command = {
|
|
"setchannelfee",
|
|
"channels",
|
|
json_setchannelfee,
|
|
"Sets specific routing fees for channel with {id} "
|
|
"(either peer ID, channel ID, short channel ID or 'all'). "
|
|
"Routing fees are defined by a fixed {base} (msat) "
|
|
"and a {ppm} (proportional per millionth) value. "
|
|
"If values for {base} or {ppm} are left out, defaults will be used. "
|
|
"{base} can also be defined in other units, for example '1sat'. "
|
|
"If {id} is 'all', the fees will be applied for all channels. ",
|
|
true /* deprecated */
|
|
};
|
|
AUTODATA(json_command, &setchannelfee_command);
|
|
|
|
static struct command_result *json_setchannel(struct command *cmd,
|
|
const char *buffer,
|
|
const jsmntok_t *obj UNNEEDED,
|
|
const jsmntok_t *params)
|
|
{
|
|
struct json_stream *response;
|
|
struct peer *peer;
|
|
struct channel **channels;
|
|
u32 *base, *ppm, *delaysecs;
|
|
struct amount_msat *htlc_min, *htlc_max;
|
|
|
|
/* Parse the JSON command */
|
|
if (!param(cmd, buffer, params,
|
|
p_req("id", param_channel_or_all, &channels),
|
|
p_opt("feebase", param_msat_u32, &base),
|
|
p_opt("feeppm", param_number, &ppm),
|
|
p_opt("htlcmin", param_msat, &htlc_min),
|
|
p_opt("htlcmax", param_msat, &htlc_max),
|
|
p_opt_def("enforcedelay", param_number, &delaysecs, 600),
|
|
NULL))
|
|
return command_param_failed();
|
|
|
|
/* Prevent obviously incorrect things! */
|
|
if (htlc_min && htlc_max
|
|
&& amount_msat_less(*htlc_max, *htlc_min)) {
|
|
return command_fail(cmd, LIGHTNINGD,
|
|
"htlcmax cannot be less than htlcmin");
|
|
}
|
|
|
|
/* Open JSON response object for later iteration */
|
|
response = json_stream_success(cmd);
|
|
json_array_start(response, "channels");
|
|
|
|
/* If the users requested 'all' channels we need to iterate */
|
|
if (channels == NULL) {
|
|
list_for_each(&cmd->ld->peers, peer, list) {
|
|
struct channel *channel;
|
|
list_for_each(&peer->channels, channel, list) {
|
|
if (channel->state != CHANNELD_NORMAL &&
|
|
channel->state != CHANNELD_AWAITING_LOCKIN &&
|
|
channel->state != DUALOPEND_AWAITING_LOCKIN)
|
|
continue;
|
|
set_channel_config(cmd, channel, base, ppm,
|
|
htlc_min, htlc_max,
|
|
*delaysecs, response, true);
|
|
}
|
|
}
|
|
/* single peer should be updated */
|
|
} else {
|
|
for (size_t i = 0; i < tal_count(channels); i++) {
|
|
set_channel_config(cmd, channels[i], base, ppm,
|
|
htlc_min, htlc_max,
|
|
*delaysecs, response, true);
|
|
}
|
|
}
|
|
|
|
/* Close and return response */
|
|
json_array_end(response);
|
|
return command_success(cmd, response);
|
|
}
|
|
|
|
static const struct json_command setchannel_command = {
|
|
"setchannel",
|
|
"channels",
|
|
json_setchannel,
|
|
"Sets fees and/or htlc_max for channel with {id} "
|
|
"(either peer ID, channel ID, short channel ID or 'all'). "
|
|
"If {feebase}, {feeppm} or {htlcmax} is missing, it is unchanged."
|
|
"{base} can also be defined in other units, for example '1sat'. "
|
|
"If {id} is 'all', the fees will be applied for all channels. "
|
|
};
|
|
AUTODATA(json_command, &setchannel_command);
|
|
|
|
#if DEVELOPER
|
|
static struct command_result *json_sign_last_tx(struct command *cmd,
|
|
const char *buffer,
|
|
const jsmntok_t *obj UNNEEDED,
|
|
const jsmntok_t *params)
|
|
{
|
|
struct node_id *peerid;
|
|
struct peer *peer;
|
|
struct json_stream *response;
|
|
struct channel *channel;
|
|
bool more_than_one;
|
|
|
|
if (!param(cmd, buffer, params,
|
|
p_req("id", param_node_id, &peerid),
|
|
NULL))
|
|
return command_param_failed();
|
|
|
|
peer = peer_by_id(cmd->ld, peerid);
|
|
if (!peer) {
|
|
return command_fail(cmd, LIGHTNINGD,
|
|
"Could not find peer with that id");
|
|
}
|
|
channel = peer_any_active_channel(peer, &more_than_one);
|
|
if (!channel || more_than_one) {
|
|
return command_fail(cmd, LIGHTNINGD,
|
|
"Could not find single active channel");
|
|
}
|
|
|
|
response = json_stream_success(cmd);
|
|
log_debug(channel->log, "dev-sign-last-tx: signing tx with %zu outputs",
|
|
channel->last_tx->wtx->num_outputs);
|
|
|
|
sign_last_tx(channel, channel->last_tx, &channel->last_sig);
|
|
json_add_tx(response, "tx", channel->last_tx);
|
|
remove_sig(channel->last_tx);
|
|
|
|
/* If we've got inflights, return them */
|
|
if (!list_empty(&channel->inflights)) {
|
|
struct channel_inflight *inflight;
|
|
|
|
json_array_start(response, "inflights");
|
|
list_for_each(&channel->inflights, inflight, list) {
|
|
sign_last_tx(channel, inflight->last_tx,
|
|
&inflight->last_sig);
|
|
json_object_start(response, NULL);
|
|
json_add_txid(response, "funding_txid",
|
|
&inflight->funding->outpoint.txid);
|
|
remove_sig(inflight->last_tx);
|
|
json_add_tx(response, "tx", channel->last_tx);
|
|
json_object_end(response);
|
|
}
|
|
json_array_end(response);
|
|
}
|
|
|
|
return command_success(cmd, response);
|
|
}
|
|
|
|
static const struct json_command dev_sign_last_tx = {
|
|
"dev-sign-last-tx",
|
|
"developer",
|
|
json_sign_last_tx,
|
|
"Sign and show the last commitment transaction with peer {id}"
|
|
};
|
|
AUTODATA(json_command, &dev_sign_last_tx);
|
|
|
|
static struct command_result *json_dev_fail(struct command *cmd,
|
|
const char *buffer,
|
|
const jsmntok_t *obj UNNEEDED,
|
|
const jsmntok_t *params)
|
|
{
|
|
struct node_id *peerid;
|
|
struct peer *peer;
|
|
struct channel *channel;
|
|
bool more_than_one;
|
|
|
|
if (!param(cmd, buffer, params,
|
|
p_req("id", param_node_id, &peerid),
|
|
NULL))
|
|
return command_param_failed();
|
|
|
|
peer = peer_by_id(cmd->ld, peerid);
|
|
if (!peer) {
|
|
return command_fail(cmd, LIGHTNINGD,
|
|
"Could not find peer with that id");
|
|
}
|
|
|
|
channel = peer_any_active_channel(peer, &more_than_one);
|
|
if (!channel || more_than_one) {
|
|
return command_fail(cmd, LIGHTNINGD,
|
|
"Could not find single active channel with peer");
|
|
}
|
|
|
|
channel_fail_permanent(channel,
|
|
REASON_USER,
|
|
"Failing due to dev-fail command");
|
|
return command_success(cmd, json_stream_success(cmd));
|
|
}
|
|
|
|
static const struct json_command dev_fail_command = {
|
|
"dev-fail",
|
|
"developer",
|
|
json_dev_fail,
|
|
"Fail with peer {id}"
|
|
};
|
|
AUTODATA(json_command, &dev_fail_command);
|
|
|
|
static void dev_reenable_commit_finished(struct subd *channeld UNUSED,
|
|
const u8 *resp UNUSED,
|
|
const int *fds UNUSED,
|
|
struct command *cmd)
|
|
{
|
|
was_pending(command_success(cmd, json_stream_success(cmd)));
|
|
}
|
|
|
|
static struct command_result *json_dev_reenable_commit(struct command *cmd,
|
|
const char *buffer,
|
|
const jsmntok_t *obj UNNEEDED,
|
|
const jsmntok_t *params)
|
|
{
|
|
struct node_id *peerid;
|
|
struct peer *peer;
|
|
u8 *msg;
|
|
struct channel *channel;
|
|
bool more_than_one;
|
|
|
|
if (!param(cmd, buffer, params,
|
|
p_req("id", param_node_id, &peerid),
|
|
NULL))
|
|
return command_param_failed();
|
|
|
|
peer = peer_by_id(cmd->ld, peerid);
|
|
if (!peer) {
|
|
return command_fail(cmd, LIGHTNINGD,
|
|
"Could not find peer with that id");
|
|
}
|
|
|
|
channel = peer_any_active_channel(peer, &more_than_one);
|
|
if (!channel || more_than_one) {
|
|
return command_fail(cmd, LIGHTNINGD,
|
|
"Peer has no active channel");
|
|
}
|
|
if (!channel->owner) {
|
|
return command_fail(cmd, LIGHTNINGD,
|
|
"Peer has no owner");
|
|
}
|
|
|
|
if (!streq(channel->owner->name, "channeld")) {
|
|
return command_fail(cmd, LIGHTNINGD,
|
|
"Peer owned by %s", channel->owner->name);
|
|
}
|
|
|
|
msg = towire_channeld_dev_reenable_commit(channel);
|
|
subd_req(peer, channel->owner, take(msg), -1, 0,
|
|
dev_reenable_commit_finished, cmd);
|
|
return command_still_pending(cmd);
|
|
}
|
|
|
|
static const struct json_command dev_reenable_commit = {
|
|
"dev-reenable-commit",
|
|
"developer",
|
|
json_dev_reenable_commit,
|
|
"Re-enable the commit timer on peer {id}"
|
|
};
|
|
AUTODATA(json_command, &dev_reenable_commit);
|
|
|
|
struct dev_forget_channel_cmd {
|
|
struct short_channel_id scid;
|
|
struct channel *channel;
|
|
bool force;
|
|
struct command *cmd;
|
|
};
|
|
|
|
static void process_dev_forget_channel(struct bitcoind *bitcoind UNUSED,
|
|
const struct bitcoin_tx_output *txout,
|
|
void *arg)
|
|
{
|
|
struct json_stream *response;
|
|
struct dev_forget_channel_cmd *forget = arg;
|
|
if (txout != NULL && !forget->force) {
|
|
was_pending(command_fail(forget->cmd, LIGHTNINGD,
|
|
"Cowardly refusing to forget channel with an "
|
|
"unspent funding output, if you know what "
|
|
"you're doing you can override with "
|
|
"`force=true`, otherwise consider `close` or "
|
|
"`dev-fail`! If you force and the channel "
|
|
"confirms we will not track the funds in the "
|
|
"channel"));
|
|
return;
|
|
}
|
|
response = json_stream_success(forget->cmd);
|
|
json_add_bool(response, "forced", forget->force);
|
|
json_add_bool(response, "funding_unspent", txout != NULL);
|
|
json_add_txid(response, "funding_txid", &forget->channel->funding.txid);
|
|
|
|
/* Set error so we don't try to reconnect. */
|
|
forget->channel->error = towire_errorfmt(forget->channel,
|
|
&forget->channel->cid,
|
|
"dev_forget_channel");
|
|
delete_channel(forget->channel);
|
|
|
|
was_pending(command_success(forget->cmd, response));
|
|
}
|
|
|
|
static struct command_result *json_dev_forget_channel(struct command *cmd,
|
|
const char *buffer,
|
|
const jsmntok_t *obj UNNEEDED,
|
|
const jsmntok_t *params)
|
|
{
|
|
struct node_id *peerid;
|
|
struct peer *peer;
|
|
struct channel *channel;
|
|
struct short_channel_id *scid;
|
|
struct channel_id *find_cid;
|
|
struct dev_forget_channel_cmd *forget = tal(cmd, struct dev_forget_channel_cmd);
|
|
forget->cmd = cmd;
|
|
|
|
bool *force;
|
|
if (!param(cmd, buffer, params,
|
|
p_req("id", param_node_id, &peerid),
|
|
p_opt("short_channel_id", param_short_channel_id, &scid),
|
|
p_opt("channel_id", param_channel_id, &find_cid),
|
|
p_opt_def("force", param_bool, &force, false),
|
|
NULL))
|
|
return command_param_failed();
|
|
|
|
forget->force = *force;
|
|
peer = peer_by_id(cmd->ld, peerid);
|
|
if (!peer) {
|
|
return command_fail(cmd, LIGHTNINGD,
|
|
"Could not find channel with that peer");
|
|
}
|
|
|
|
forget->channel = NULL;
|
|
list_for_each(&peer->channels, channel, list) {
|
|
/* Check for channel id first */
|
|
if (find_cid) {
|
|
if (!channel_id_eq(find_cid, &channel->cid))
|
|
continue;
|
|
}
|
|
if (scid) {
|
|
if (!channel->scid)
|
|
continue;
|
|
if (!short_channel_id_eq(channel->scid, scid))
|
|
continue;
|
|
}
|
|
if (forget->channel) {
|
|
return command_fail(cmd, LIGHTNINGD,
|
|
"Multiple channels:"
|
|
" please specify short_channel_id");
|
|
}
|
|
forget->channel = channel;
|
|
}
|
|
if (!forget->channel) {
|
|
return command_fail(cmd, LIGHTNINGD,
|
|
"No channels matching that peer_id%s",
|
|
scid ? " and that short_channel_id" : "");
|
|
}
|
|
|
|
if (channel_has_htlc_out(forget->channel) ||
|
|
channel_has_htlc_in(forget->channel)) {
|
|
return command_fail(cmd, LIGHTNINGD,
|
|
"This channel has HTLCs attached and it is "
|
|
"not safe to forget it. Please use `close` "
|
|
"or `dev-fail` instead.");
|
|
}
|
|
|
|
if (!channel_unsaved(forget->channel))
|
|
bitcoind_getutxout(cmd->ld->topology->bitcoind,
|
|
&forget->channel->funding,
|
|
process_dev_forget_channel, forget);
|
|
return command_still_pending(cmd);
|
|
}
|
|
|
|
static const struct json_command dev_forget_channel_command = {
|
|
"dev-forget-channel",
|
|
"developer",
|
|
json_dev_forget_channel,
|
|
"Forget the channel with peer {id}, ignore UTXO check with {force}='true'.", false,
|
|
"Forget the channel with peer {id}. Checks if the channel is still active by checking its funding transaction. Check can be ignored by setting {force} to 'true'"
|
|
};
|
|
AUTODATA(json_command, &dev_forget_channel_command);
|
|
|
|
static void channeld_memleak_req_done(struct subd *channeld,
|
|
const u8 *msg, const int *fds UNUSED,
|
|
struct leak_detect *leaks)
|
|
{
|
|
bool found_leak;
|
|
|
|
if (!fromwire_channeld_dev_memleak_reply(msg, &found_leak))
|
|
fatal("Bad channel_dev_memleak");
|
|
|
|
if (found_leak)
|
|
report_subd_memleak(leaks, channeld);
|
|
}
|
|
|
|
static void onchaind_memleak_req_done(struct subd *onchaind,
|
|
const u8 *msg, const int *fds UNUSED,
|
|
struct leak_detect *leaks)
|
|
{
|
|
bool found_leak;
|
|
|
|
if (!fromwire_onchaind_dev_memleak_reply(msg, &found_leak))
|
|
fatal("Bad onchaind_dev_memleak");
|
|
|
|
if (found_leak)
|
|
report_subd_memleak(leaks, onchaind);
|
|
}
|
|
|
|
static void openingd_memleak_req_done(struct subd *open_daemon,
|
|
const u8 *msg, const int *fds UNUSED,
|
|
struct leak_detect *leaks)
|
|
{
|
|
bool found_leak;
|
|
|
|
if (!fromwire_openingd_dev_memleak_reply(msg, &found_leak))
|
|
fatal("Bad opening_dev_memleak");
|
|
|
|
if (found_leak)
|
|
report_subd_memleak(leaks, open_daemon);
|
|
}
|
|
|
|
static void dualopend_memleak_req_done(struct subd *dualopend,
|
|
const u8 *msg, const int *fds UNUSED,
|
|
struct leak_detect *leaks)
|
|
{
|
|
bool found_leak;
|
|
|
|
if (!fromwire_dualopend_dev_memleak_reply(msg, &found_leak))
|
|
fatal("Bad dualopend_dev_memleak");
|
|
|
|
if (found_leak)
|
|
report_subd_memleak(leaks, dualopend);
|
|
}
|
|
|
|
void peer_dev_memleak(struct lightningd *ld, struct leak_detect *leaks)
|
|
{
|
|
struct peer *p;
|
|
|
|
list_for_each(&ld->peers, p, list) {
|
|
struct channel *c;
|
|
if (p->uncommitted_channel) {
|
|
struct subd *openingd = p->uncommitted_channel->open_daemon;
|
|
start_leak_request(subd_req(openingd, openingd,
|
|
take(towire_openingd_dev_memleak(NULL)),
|
|
-1, 0, openingd_memleak_req_done, leaks),
|
|
leaks);
|
|
}
|
|
|
|
list_for_each(&p->channels, c, list) {
|
|
if (!c->owner)
|
|
continue;
|
|
if (streq(c->owner->name, "channeld")) {
|
|
start_leak_request(subd_req(c, c->owner,
|
|
take(towire_channeld_dev_memleak(NULL)),
|
|
-1, 0, channeld_memleak_req_done, leaks),
|
|
leaks);
|
|
} else if (streq(c->owner->name, "onchaind")) {
|
|
start_leak_request(subd_req(c, c->owner,
|
|
take(towire_onchaind_dev_memleak(NULL)),
|
|
-1, 0, onchaind_memleak_req_done, leaks),
|
|
leaks);
|
|
} else if (streq(c->owner->name, "dualopend")) {
|
|
start_leak_request(subd_req(c, c->owner,
|
|
take(towire_dualopend_dev_memleak(NULL)),
|
|
-1, 0, dualopend_memleak_req_done, leaks),
|
|
leaks);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
#endif /* DEVELOPER */
|