2017-01-14 20:32:32 +01:00
from binascii import hexlify , unhexlify
from concurrent import futures
2017-06-20 16:41:49 +02:00
from decimal import Decimal
2017-01-14 20:32:32 +01:00
from hashlib import sha256
2017-08-28 17:53:01 +02:00
from lightning import LightningRpc
2017-01-14 20:32:32 +01:00
2017-04-29 03:02:27 +02:00
import copy
import json
2017-01-14 20:32:32 +01:00
import logging
2017-09-26 06:57:19 +02:00
import queue
2017-01-14 20:32:32 +01:00
import os
2017-07-05 07:33:40 +02:00
import random
2017-08-17 03:02:55 +02:00
import re
2017-06-20 16:41:49 +02:00
import sqlite3
2017-07-05 07:33:40 +02:00
import string
import sys
2017-01-14 20:32:32 +01:00
import tempfile
2017-08-18 06:43:52 +02:00
import threading
2017-01-14 20:32:32 +01:00
import time
import unittest
2017-01-17 23:30:22 +01:00
import utils
2017-01-14 20:32:32 +01:00
bitcoind = None
TEST_DIR = tempfile . mkdtemp ( prefix = ' lightning- ' )
2017-08-31 04:02:47 +02:00
VALGRIND = os . getenv ( " NO_VALGRIND " , " 0 " ) == " 0 "
2017-10-24 04:06:14 +02:00
DEVELOPER = os . getenv ( " DEVELOPER " , " 0 " ) == " 1 "
2017-05-02 14:32:12 +02:00
TEST_DEBUG = os . getenv ( " TEST_DEBUG " , " 0 " ) == " 1 "
2017-01-14 20:32:32 +01:00
2017-05-02 14:32:12 +02:00
print ( " Testing results are in {} " . format ( TEST_DIR ) )
if TEST_DEBUG :
2017-01-14 20:32:32 +01:00
logging . basicConfig ( level = logging . DEBUG , stream = sys . stdout )
logging . info ( " Tests running in ' %s ' " , TEST_DIR )
2017-04-29 03:02:27 +02:00
def to_json ( arg ) :
return json . loads ( json . dumps ( arg ) )
2017-01-14 20:32:32 +01:00
def setupBitcoind ( ) :
global bitcoind
2017-01-17 23:30:22 +01:00
bitcoind = utils . BitcoinD ( rpcport = 28332 )
2017-01-14 20:32:32 +01:00
bitcoind . start ( )
2017-11-28 05:39:22 +01:00
info = bitcoind . rpc . getblockchaininfo ( )
2017-01-14 20:32:32 +01:00
# Make sure we have segwit and some funds
if info [ ' blocks ' ] < 432 :
logging . debug ( " SegWit not active, generating some more blocks " )
2017-11-28 05:39:22 +01:00
bitcoind . generate_block ( 432 - info [ ' blocks ' ] )
elif bitcoind . rpc . getwalletinfo ( ) [ ' balance ' ] < 1 :
2017-01-14 20:32:32 +01:00
logging . debug ( " Insufficient balance, generating 1 block " )
2017-11-28 05:39:22 +01:00
bitcoind . generate_block ( 1 )
2017-01-14 20:32:32 +01:00
2017-07-04 20:36:24 +02:00
2017-08-17 15:30:24 +02:00
def wait_for ( success , timeout = 30 , interval = 0.1 ) :
2017-07-04 20:36:24 +02:00
start_time = time . time ( )
while not success ( ) and time . time ( ) < start_time + timeout :
2017-08-17 15:30:24 +02:00
time . sleep ( interval )
2017-07-04 20:36:24 +02:00
if time . time ( ) > start_time + timeout :
raise ValueError ( " Error waiting for {} " , success )
2017-07-04 20:47:50 +02:00
def sync_blockheight ( nodes ) :
target = bitcoind . rpc . getblockcount ( )
for n in nodes :
wait_for ( lambda : n . rpc . getinfo ( ) [ ' blockheight ' ] == target )
2017-01-14 20:32:32 +01:00
def tearDownBitcoind ( ) :
global bitcoind
try :
bitcoind . rpc . stop ( )
except :
bitcoind . proc . kill ( )
bitcoind . proc . wait ( )
def setUpModule ( ) :
setupBitcoind ( )
def tearDownModule ( ) :
tearDownBitcoind ( )
2017-09-30 15:23:03 +02:00
def breakpoint ( ) :
import pdb ; pdb . set_trace ( )
2017-01-14 20:32:32 +01:00
class NodeFactory ( object ) :
""" A factory to setup and start `lightningd` daemons.
"""
2017-11-20 16:16:17 +01:00
def __init__ ( self , testname , bitcoind , executor ) :
self . testname = testname
2017-01-14 20:32:32 +01:00
self . next_id = 1
self . nodes = [ ]
self . executor = executor
2017-11-20 16:16:17 +01:00
self . bitcoind = bitcoind
2017-01-14 20:32:32 +01:00
2017-12-02 07:36:15 +01:00
def get_node ( self , disconnect = None , options = None , may_fail = False , random_hsm = False ) :
2017-01-14 20:32:32 +01:00
node_id = self . next_id
self . next_id + = 1
lightning_dir = os . path . join (
2017-11-20 16:16:17 +01:00
TEST_DIR , self . testname , " lightning- {} / " . format ( node_id ) )
2017-01-14 20:32:32 +01:00
socket_path = os . path . join ( lightning_dir , " lightning-rpc " ) . format ( node_id )
2017-01-17 23:30:22 +01:00
port = 16330 + node_id
2017-12-02 07:36:15 +01:00
daemon = utils . LightningD ( lightning_dir , self . bitcoind . bitcoin_dir , port = port , random_hsm = random_hsm )
2017-08-28 17:53:01 +02:00
# If we have a disconnect string, dump it to a file for daemon.
if disconnect :
with open ( os . path . join ( lightning_dir , " dev_disconnect " ) , " w " ) as f :
f . write ( " \n " . join ( disconnect ) )
daemon . cmd_line . append ( " --dev-disconnect=dev_disconnect " )
2017-10-24 04:06:14 +02:00
if DEVELOPER :
daemon . cmd_line . append ( " --dev-fail-on-subdaemon-fail " )
2017-10-31 01:04:59 +01:00
if VALGRIND :
daemon . cmd_line . append ( " --dev-no-backtrace " )
2017-10-09 03:30:00 +02:00
opts = [ ] if options is None else options
for opt in opts :
daemon . cmd_line . append ( opt )
2017-08-28 17:53:01 +02:00
rpc = LightningRpc ( socket_path , self . executor )
2017-01-17 23:30:22 +01:00
2017-11-20 16:16:17 +01:00
node = utils . LightningNode ( daemon , rpc , self . bitcoind , self . executor , may_fail = may_fail )
2017-01-14 20:32:32 +01:00
self . nodes . append ( node )
if VALGRIND :
node . daemon . cmd_line = [
2017-01-24 05:07:56 +01:00
' valgrind ' ,
2017-01-14 20:32:32 +01:00
' -q ' ,
2017-05-05 09:24:38 +02:00
' --trace-children=yes ' ,
' --trace-children-skip=*bitcoin-cli* ' ,
2017-01-14 20:32:32 +01:00
' --error-exitcode=7 ' ,
2017-08-17 03:02:55 +02:00
' --log-file= {} /valgrind-errors. % p ' . format ( node . daemon . lightning_dir )
2017-01-14 20:32:32 +01:00
] + node . daemon . cmd_line
node . daemon . start ( )
# Cache `getinfo`, we'll be using it a lot
node . info = node . rpc . getinfo ( )
return node
def killall ( self ) :
2017-09-28 17:52:48 +02:00
rcs = [ ]
failed = False
2017-01-14 20:32:32 +01:00
for n in self . nodes :
2017-09-28 17:52:48 +02:00
try :
n . stop ( )
except :
failed = True
rcs . append ( n . daemon . proc . returncode )
2017-11-09 17:07:18 +01:00
return rcs
2017-01-14 20:32:32 +01:00
2017-01-17 23:30:22 +01:00
class BaseLightningDTests ( unittest . TestCase ) :
2017-01-14 20:32:32 +01:00
def setUp ( self ) :
# Most of the executor threads will be waiting for IO, so
# let's have a few of them
self . executor = futures . ThreadPoolExecutor ( max_workers = 20 )
2017-11-20 16:16:17 +01:00
self . node_factory = NodeFactory ( self . _testMethodName , bitcoind , self . executor )
2017-01-14 20:32:32 +01:00
2017-05-09 13:48:16 +02:00
def getValgrindErrors ( self , node ) :
2017-08-17 03:02:55 +02:00
for error_file in os . listdir ( node . daemon . lightning_dir ) :
2017-10-31 01:05:08 +01:00
if not re . fullmatch ( " valgrind-errors. \ d+ " , error_file ) :
2017-08-17 03:02:55 +02:00
continue ;
with open ( os . path . join ( node . daemon . lightning_dir , error_file ) , ' r ' ) as f :
errors = f . read ( ) . strip ( )
if errors :
return errors , error_file
return None , None
2017-05-09 13:48:16 +02:00
def printValgrindErrors ( self , node ) :
errors , fname = self . getValgrindErrors ( node )
if errors :
print ( " - " * 31 , " Valgrind errors " , " - " * 32 )
print ( " Valgrind error file: " , fname )
print ( errors )
print ( " - " * 80 )
return 1 if errors else 0
2017-09-12 06:55:54 +02:00
def getCrashLog ( self , node ) :
2017-09-28 18:01:09 +02:00
if node . may_fail :
2017-09-20 06:45:41 +02:00
return None , None
2017-09-12 06:55:54 +02:00
try :
crashlog = os . path . join ( node . daemon . lightning_dir , ' crash.log ' )
with open ( crashlog , ' r ' ) as f :
return f . readlines ( ) , crashlog
except :
return None , None
def printCrashLog ( self , node ) :
errors , fname = self . getCrashLog ( node )
if errors :
print ( " - " * 10 , " {} (last 50 lines) " . format ( fname ) , " - " * 10 )
for l in errors [ - 50 : ] :
print ( l , end = ' ' )
print ( " - " * 80 )
return 1 if errors else 0
2017-01-14 20:32:32 +01:00
def tearDown ( self ) :
2017-11-09 17:07:18 +01:00
rcs = self . node_factory . killall ( )
2017-01-14 20:32:32 +01:00
self . executor . shutdown ( wait = False )
2017-05-12 12:08:14 +02:00
2017-09-12 06:55:54 +02:00
err_count = 0
2017-05-12 12:08:14 +02:00
# Do not check for valgrind error files if it is disabled
if VALGRIND :
for node in self . node_factory . nodes :
err_count + = self . printValgrindErrors ( node )
if err_count :
2017-11-09 17:07:18 +01:00
raise ValueError ( " {} nodes reported valgrind errors " . format ( err_count ) )
2017-01-17 23:30:22 +01:00
2017-09-12 06:55:54 +02:00
for node in self . node_factory . nodes :
err_count + = self . printCrashLog ( node )
if err_count :
2017-11-09 17:07:18 +01:00
raise ValueError ( " {} nodes had crash.log files " . format ( err_count ) )
# Which nodes may fail? Mask away the ones that we know will fail
failmask = [ not n . may_fail for n in self . node_factory . nodes ]
unexpected = [ ( failmask [ i ] * rcs [ i ] ) for i in range ( len ( rcs ) ) ]
if len ( [ u for u in unexpected if u > 0 ] ) > 0 :
raise Exception ( " At least one lightning exited with unexpected non-zero return code: {} " . format ( unexpected ) )
2017-09-12 06:55:54 +02:00
2017-01-17 23:30:22 +01:00
class LightningDTests ( BaseLightningDTests ) :
2017-04-29 03:02:26 +02:00
def connect ( self ) :
2017-08-28 17:53:01 +02:00
l1 = self . node_factory . get_node ( )
l2 = self . node_factory . get_node ( )
2017-12-04 10:12:22 +01:00
ret = l1 . rpc . connect ( l2 . info [ ' id ' ] , ' localhost ' , l2 . info [ ' port ' ] )
2017-01-17 23:30:22 +01:00
assert ret [ ' id ' ] == l2 . info [ ' id ' ]
gossipd: rewrite to do the handshake internally.
Now the flow is much simpler from a lightningd POV:
1. If we want to connect to a peer, just send gossipd `gossipctl_reach_peer`.
2. Every new peer, gossipd hands up to lightningd, with global/local features
and the peer fd and a gossip fd using `gossip_peer_connected`
3. If lightningd doesn't want it, it just hands the peerfd and global/local
features back to gossipd using `gossipctl_handle_peer`
4. If a peer sends a non-gossip msg (eg `open_channel`) the gossipd sends
it up using `gossip_peer_nongossip`.
5. If lightningd wants to fund a channel, it simply calls `release_channel`.
Notes:
* There's no more "unique_id": we use the peer id.
* For the moment, we don't ask gossipd when we're told to list peers, so
connected peers without a channel don't appear in the JSON getpeers API.
* We add a `gossipctl_peer_addrhint` for the moment, so you can connect to
a specific ip/port, but using other sources is a TODO.
* We now (correctly) only give up on reaching a peer after we exchange init
messages, which changes the test_disconnect case.
Signed-off-by: Rusty Russell <rusty@rustcorp.com.au>
2017-10-11 12:09:49 +02:00
l1 . daemon . wait_for_log ( ' WIRE_GOSSIPCTL_HANDLE_PEER ' )
l2 . daemon . wait_for_log ( ' WIRE_GOSSIPCTL_HANDLE_PEER ' )
2017-04-29 03:02:26 +02:00
return l1 , l2
2017-02-08 16:56:39 +01:00
2017-10-22 15:32:46 +02:00
# Returns the short channel-id: <blocknum>:<txnum>:<outnum>
2017-07-05 07:33:40 +02:00
def fund_channel ( self , l1 , l2 , amount ) :
2017-04-12 20:08:48 +02:00
addr = l1 . rpc . newaddr ( ) [ ' address ' ]
2017-03-15 21:52:49 +01:00
2017-04-29 03:02:26 +02:00
txid = l1 . bitcoin . rpc . sendtoaddress ( addr , amount / 10 * * 8 + 0.01 )
2017-03-15 21:52:49 +01:00
tx = l1 . bitcoin . rpc . getrawtransaction ( txid )
l1 . rpc . addfunds ( tx )
2017-10-22 15:32:46 +02:00
# Generate a block, so we know next tx will be first in block.
2017-11-28 05:39:22 +01:00
l1 . bitcoin . generate_block ( 1 )
2017-10-22 15:32:46 +02:00
tx = l1 . rpc . fundchannel ( l2 . info [ ' id ' ] , amount ) [ ' tx ' ]
2017-05-05 08:41:46 +02:00
# Technically, this is async to fundchannel.
l1 . daemon . wait_for_log ( ' sendrawtx exit 0 ' )
2017-11-28 05:39:22 +01:00
l1 . bitcoin . generate_block ( 1 )
2017-05-22 13:27:20 +02:00
l1 . daemon . wait_for_log ( ' -> CHANNELD_NORMAL ' )
l2 . daemon . wait_for_log ( ' -> CHANNELD_NORMAL ' )
2017-03-15 21:52:49 +01:00
2017-10-22 15:32:46 +02:00
# Hacky way to find our output.
2017-10-31 03:06:36 +01:00
decoded = bitcoind . rpc . decoderawtransaction ( tx )
for out in decoded [ ' vout ' ] :
test_lightning.py: fix float insanity with values.
When is 0.01 != 0.01? When there are floats involved! Jenkins hit an
error once, I have no idea why.
This works around the following intermittant error:
ERROR: test_closing_negotiation_reconnect (__main__.LightningDTests)
----------------------------------------------------------------------
Traceback (most recent call last):
File "tests/test_lightningd.py", line 1601, in test_closing_negotiation_reconnect
self.fund_channel(l1, l2, 10**6)
File "tests/test_lightningd.py", line 241, in fund_channel
raise ValueError("Can't find {} payment in {}".format(amount, tx))
ValueError: Can't find 1000000 payment in 02000000000101b0b27be92916faee59f197c263e1ae7b44c0e59acdf2c385c55b5b04670ac026010000001716001401fad90abcd66697e2592164722de4a95ebee165ffffffff02651d0f0000000000160014c2ccab171c2a5be9dab52ec41b825863024c546640420f00000000002200205b8cd3b914cf67cdd8fa6273c930353dd36476734fbd962102c2df53b90880cd02473044022017bd19a0ee85532f67a280c71ed02d1321f08975334bd281527478022265225702202ec448bf9c0890a31a26f0ef4f03d298c8ec16b277faff09a70ddd335df44b6e012103d745445c9362665f22e0d96e9e766f273f3260dea39c8a76bfa05dd2684ddccf00000000
For testing, that tx decodes to:
{
"txid": "0165e92be762b352b665b76b9872d5189e1b2a8faf4918ab3cca7cd5d4b6a5fa",
"hash": "8af9a36c79ee5243468c5cbed1c80f10238fba405f0ad957a0c2cfc46fb632f5",
"version": 2,
"size": 257,
"vsize": 176,
"locktime": 0,
"vin": [
{
"txid": "26c00a67045b5bc585c3f2cd9ae5c0447baee163c297f159eefa1629e97bb2b0",
"vout": 1,
"scriptSig": {
"asm": "001401fad90abcd66697e2592164722de4a95ebee165",
"hex": "16001401fad90abcd66697e2592164722de4a95ebee165"
},
"txinwitness": [
"3044022017bd19a0ee85532f67a280c71ed02d1321f08975334bd281527478022265225702202ec448bf9c0890a31a26f0ef4f03d298c8ec16b277faff09a70ddd335df44b6e01",
"03d745445c9362665f22e0d96e9e766f273f3260dea39c8a76bfa05dd2684ddccf"
],
"sequence": 4294967295
}
],
"vout": [
{
"value": 0.00990565,
"n": 0,
"scriptPubKey": {
"asm": "0 c2ccab171c2a5be9dab52ec41b825863024c5466",
"hex": "0014c2ccab171c2a5be9dab52ec41b825863024c5466",
"type": "witness_v0_keyhash"
}
},
{
"value": 0.01000000,
"n": 1,
"scriptPubKey": {
"asm": "0 5b8cd3b914cf67cdd8fa6273c930353dd36476734fbd962102c2df53b90880cd",
"hex": "00205b8cd3b914cf67cdd8fa6273c930353dd36476734fbd962102c2df53b90880cd",
"type": "witness_v0_scripthash"
}
}
]
}
Signed-off-by: Rusty Russell <rusty@rustcorp.com.au>
2017-10-22 22:32:10 +02:00
# Sometimes a float? Sometimes a decimal? WTF Python?!
if out [ ' scriptPubKey ' ] [ ' type ' ] == ' witness_v0_scripthash ' :
if out [ ' value ' ] == Decimal ( amount ) / 10 * * 8 or out [ ' value ' ] * 10 * * 8 == amount :
return " {} :1: {} " . format ( bitcoind . rpc . getblockcount ( ) , out [ ' n ' ] )
2017-10-31 03:06:36 +01:00
# Intermittant decoding failure. See if it decodes badly twice?
decoded2 = bitcoind . rpc . decoderawtransaction ( tx )
raise ValueError ( " Can ' t find {} payment in {} (1= {} 2= {} ) " . format ( amount , tx , decoded , decoded2 ) )
2017-10-22 15:32:46 +02:00
2017-11-17 15:09:38 +01:00
def line_graph ( self , n = 2 ) :
""" Build a line graph of the specified length and fund it.
"""
nodes = [ self . node_factory . get_node ( ) for _ in range ( n ) ]
for i in range ( len ( nodes ) - 1 ) :
nodes [ i ] . rpc . connect (
nodes [ i + 1 ] . info [ ' id ' ] ,
2017-12-04 10:12:22 +01:00
' localhost ' ,
nodes [ i + 1 ] . info [ ' port ' ]
2017-11-17 15:09:38 +01:00
)
self . fund_channel ( nodes [ i ] , nodes [ i + 1 ] , 10 * * 6 )
return nodes
2017-08-18 06:43:52 +02:00
def pay ( self , lsrc , ldst , amt , label = None , async = False ) :
2017-07-05 07:33:40 +02:00
if not label :
label = ' ' . join ( random . choice ( string . ascii_letters + string . digits ) for _ in range ( 20 ) )
2017-10-26 05:06:19 +02:00
rhash = ldst . rpc . invoice ( amt , label , label ) [ ' rhash ' ]
2017-07-04 02:48:43 +02:00
assert ldst . rpc . listinvoice ( label ) [ 0 ] [ ' complete ' ] == False
2017-09-07 21:33:55 +02:00
routestep = {
' msatoshi ' : amt ,
' id ' : ldst . info [ ' id ' ] ,
' delay ' : 5 ,
' channel ' : ' 1:1:1 '
}
2017-08-18 06:43:52 +02:00
def wait_pay ( ) :
# Up to 10 seconds for payment to succeed.
start_time = time . time ( )
while not ldst . rpc . listinvoice ( label ) [ 0 ] [ ' complete ' ] :
if time . time ( ) > start_time + 10 :
raise TimeoutError ( ' Payment timed out ' )
time . sleep ( 0.1 )
if async :
2017-09-07 21:33:55 +02:00
self . executor . submit ( lsrc . rpc . sendpay , to_json ( [ routestep ] ) , rhash , async = False )
2017-08-18 06:43:52 +02:00
return self . executor . submit ( wait_pay )
else :
2017-09-07 21:33:55 +02:00
lsrc . rpc . sendpay ( to_json ( [ routestep ] ) , rhash , async = False )
2017-07-04 02:48:43 +02:00
2017-09-12 06:55:54 +02:00
def test_shutdown ( self ) :
l1 = self . node_factory . get_node ( )
l1 . rpc . stop ( )
2017-10-26 05:06:19 +02:00
def test_invoice ( self ) :
l1 = self . node_factory . get_node ( )
before = int ( time . time ( ) )
inv = l1 . rpc . invoice ( 123000 , ' label ' , ' description ' )
after = int ( time . time ( ) )
b11 = l1 . rpc . decodepay ( inv [ ' bolt11 ' ] )
assert b11 [ ' currency ' ] == ' tb '
assert b11 [ ' timestamp ' ] > = before
assert b11 [ ' timestamp ' ] < = after
assert b11 [ ' payment_hash ' ] == inv [ ' rhash ' ]
assert b11 [ ' description ' ] == ' description '
assert b11 [ ' expiry ' ] == 3600
assert b11 [ ' payee ' ] == l1 . info [ ' id ' ]
2017-11-08 14:34:11 +01:00
2017-11-21 07:27:28 +01:00
def test_invoice_expiry ( self ) :
l1 , l2 = self . connect ( )
chanid = self . fund_channel ( l1 , l2 , 10 * * 6 )
# Wait for route propagation.
2017-11-28 05:39:22 +01:00
bitcoind . generate_block ( 5 )
2017-11-21 07:27:28 +01:00
l1 . daemon . wait_for_logs ( [ ' Received channel_update for channel {} \ (0 \ ) '
. format ( chanid ) ,
' Received channel_update for channel {} \ (1 \ ) '
. format ( chanid ) ] )
inv = l2 . rpc . invoice ( 123000 , ' test_pay ' , ' description ' , 1 ) [ ' bolt11 ' ]
time . sleep ( 2 )
self . assertRaises ( ValueError , l1 . rpc . pay , inv )
assert l2 . rpc . listinvoice ( ' test_pay ' ) [ 0 ] [ ' complete ' ] == False
assert l2 . rpc . listinvoice ( ' test_pay ' ) [ 0 ] [ ' expiry_time ' ] < time . time ( )
2017-04-29 03:02:26 +02:00
def test_connect ( self ) :
l1 , l2 = self . connect ( )
2017-10-23 06:19:38 +02:00
# These should be in gossipd.
assert l1 . rpc . getpeer ( l2 . info [ ' id ' ] ) [ ' state ' ] == ' GOSSIPING '
assert l2 . rpc . getpeer ( l1 . info [ ' id ' ] ) [ ' state ' ] == ' GOSSIPING '
2017-04-29 03:02:27 +02:00
gossipd: rewrite to do the handshake internally.
Now the flow is much simpler from a lightningd POV:
1. If we want to connect to a peer, just send gossipd `gossipctl_reach_peer`.
2. Every new peer, gossipd hands up to lightningd, with global/local features
and the peer fd and a gossip fd using `gossip_peer_connected`
3. If lightningd doesn't want it, it just hands the peerfd and global/local
features back to gossipd using `gossipctl_handle_peer`
4. If a peer sends a non-gossip msg (eg `open_channel`) the gossipd sends
it up using `gossip_peer_nongossip`.
5. If lightningd wants to fund a channel, it simply calls `release_channel`.
Notes:
* There's no more "unique_id": we use the peer id.
* For the moment, we don't ask gossipd when we're told to list peers, so
connected peers without a channel don't appear in the JSON getpeers API.
* We add a `gossipctl_peer_addrhint` for the moment, so you can connect to
a specific ip/port, but using other sources is a TODO.
* We now (correctly) only give up on reaching a peer after we exchange init
messages, which changes the test_disconnect case.
Signed-off-by: Rusty Russell <rusty@rustcorp.com.au>
2017-10-11 12:09:49 +02:00
# Both gossipds will have them as new peers once handed back.
l1 . daemon . wait_for_log ( ' handle_peer {} : new peer ' . format ( l2 . info [ ' id ' ] ) )
l2 . daemon . wait_for_log ( ' handle_peer {} : new peer ' . format ( l1 . info [ ' id ' ] ) )
2017-04-29 03:02:27 +02:00
2017-06-20 07:59:03 +02:00
def test_balance ( self ) :
l1 , l2 = self . connect ( )
self . fund_channel ( l1 , l2 , 10 * * 6 )
p1 = l1 . rpc . getpeer ( l2 . info [ ' id ' ] , ' info ' )
p2 = l2 . rpc . getpeer ( l1 . info [ ' id ' ] , ' info ' )
assert p1 [ ' msatoshi_to_us ' ] == 10 * * 6 * 1000
assert p1 [ ' msatoshi_total ' ] == 10 * * 6 * 1000
assert p2 [ ' msatoshi_to_us ' ] == 0
assert p2 [ ' msatoshi_total ' ] == 10 * * 6 * 1000
2017-10-26 05:05:19 +02:00
def test_decodepay ( self ) :
l1 = self . node_factory . get_node ( )
# BOLT #11:
# > ### Please make a donation of any amount using payment_hash 0001020304050607080900010203040506070809000102030405060708090102 to me @03e7156ae33b0a208d0744199163177e909e80176e55d97a2f221ede0f934dd9ad
# > lnbc1pvjluezpp5qqqsyqcyq5rqwzqfqqqsyqcyq5rqwzqfqqqsyqcyq5rqwzqfqypqdpl2pkx2ctnv5sxxmmwwd5kgetjypeh2ursdae8g6twvus8g6rfwvs8qun0dfjkxaq8rkx3yf5tcsyz3d73gafnh3cax9rn449d9p5uxz9ezhhypd0elx87sjle52x86fux2ypatgddc6k63n7erqz25le42c4u4ecky03ylcqca784w
2017-11-08 14:34:11 +01:00
#
2017-10-26 05:05:19 +02:00
# Breakdown:
2017-11-08 14:34:11 +01:00
#
2017-10-26 05:05:19 +02:00
# * `lnbc`: prefix, lightning on bitcoin mainnet
# * `1`: Bech32 separator
# * `pvjluez`: timestamp (1496314658)
# * `p`: payment hash
# * `p5`: `data_length` (`p` = 1, `5` = 20. 1 * 32 + 20 == 52)
# * `qqqsyqcyq5rqwzqfqqqsyqcyq5rqwzqfqqqsyqcyq5rqwzqfqypq`: payment hash 0001020304050607080900010203040506070809000102030405060708090102
# * `d`: short description
# * `pl`: `data_length` (`p` = 1, `l` = 31. 1 * 32 + 31 == 63)
# * `2pkx2ctnv5sxxmmwwd5kgetjypeh2ursdae8g6twvus8g6rfwvs8qun0dfjkxaq`: 'Please consider supporting this project'
# * `32vjcgqxyuj7nqphl3xmmhls2rkl3t97uan4j0xa87gj5779czc8p0z58zf5wpt9ggem6adl64cvawcxlef9djqwp2jzzfvs272504sp`: signature
# * `0lkg3c`: Bech32 checksum
b11 = l1 . rpc . decodepay ( ' lnbc1pvjluezpp5qqqsyqcyq5rqwzqfqqqsyqcyq5rqwzqfqqqsyqcyq5rqwzqfqypqdpl2pkx2ctnv5sxxmmwwd5kgetjypeh2ursdae8g6twvus8g6rfwvs8qun0dfjkxaq8rkx3yf5tcsyz3d73gafnh3cax9rn449d9p5uxz9ezhhypd0elx87sjle52x86fux2ypatgddc6k63n7erqz25le42c4u4ecky03ylcqca784w ' )
assert b11 [ ' currency ' ] == ' bc '
assert b11 [ ' timestamp ' ] == 1496314658
assert b11 [ ' payment_hash ' ] == ' 0001020304050607080900010203040506070809000102030405060708090102 '
assert b11 [ ' description ' ] == ' Please consider supporting this project '
assert b11 [ ' expiry ' ] == 3600
assert b11 [ ' payee ' ] == ' 03e7156ae33b0a208d0744199163177e909e80176e55d97a2f221ede0f934dd9ad '
2017-11-08 14:34:11 +01:00
2017-10-26 05:05:19 +02:00
# BOLT #11:
# > ### Please send $3 for a cup of coffee to the same peer, within 1 minute
# > lnbc2500u1pvjluezpp5qqqsyqcyq5rqwzqfqqqsyqcyq5rqwzqfqqqsyqcyq5rqwzqfqypqdq5xysxxatsyp3k7enxv4jsxqzpuaztrnwngzn3kdzw5hydlzf03qdgm2hdq27cqv3agm2awhz5se903vruatfhq77w3ls4evs3ch9zw97j25emudupq63nyw24cg27h2rspfj9srp
2017-11-08 14:34:11 +01:00
#
2017-10-26 05:05:19 +02:00
# Breakdown:
2017-11-08 14:34:11 +01:00
#
2017-10-26 05:05:19 +02:00
# * `lnbc`: prefix, lightning on bitcoin mainnet
# * `2500u`: amount (2500 micro-bitcoin)
# * `1`: Bech32 separator
# * `pvjluez`: timestamp (1496314658)
# * `p`: payment hash...
# * `d`: short description
# * `q5`: `data_length` (`q` = 0, `5` = 20. 0 * 32 + 20 == 20)
# * `xysxxatsyp3k7enxv4js`: '1 cup coffee'
# * `x`: expiry time
# * `qz`: `data_length` (`q` = 0, `z` = 2. 0 * 32 + 2 == 2)
# * `pu`: 60 seconds (`p` = 1, `u` = 28. 1 * 32 + 28 == 60)
# * `azh8qt5w7qeewkmxtv55khqxvdfs9zzradsvj7rcej9knpzdwjykcq8gv4v2dl705pjadhpsc967zhzdpuwn5qzjm0s4hqm2u0vuhhqq`: signature
# * `7vc09u`: Bech32 checksum
b11 = l1 . rpc . decodepay ( ' lnbc2500u1pvjluezpp5qqqsyqcyq5rqwzqfqqqsyqcyq5rqwzqfqqqsyqcyq5rqwzqfqypqdq5xysxxatsyp3k7enxv4jsxqzpuaztrnwngzn3kdzw5hydlzf03qdgm2hdq27cqv3agm2awhz5se903vruatfhq77w3ls4evs3ch9zw97j25emudupq63nyw24cg27h2rspfj9srp ' )
assert b11 [ ' currency ' ] == ' bc '
assert b11 [ ' msatoshi ' ] == 2500 * 10 * * 11 / / 1000000
assert b11 [ ' timestamp ' ] == 1496314658
assert b11 [ ' payment_hash ' ] == ' 0001020304050607080900010203040506070809000102030405060708090102 '
assert b11 [ ' description ' ] == ' 1 cup coffee '
assert b11 [ ' expiry ' ] == 60
assert b11 [ ' payee ' ] == ' 03e7156ae33b0a208d0744199163177e909e80176e55d97a2f221ede0f934dd9ad '
# BOLT #11:
# > ### Now send $24 for an entire list of things (hashed)
# > lnbc20m1pvjluezpp5qqqsyqcyq5rqwzqfqqqsyqcyq5rqwzqfqqqsyqcyq5rqwzqfqypqhp58yjmdan79s6qqdhdzgynm4zwqd5d7xmw5fk98klysy043l2ahrqscc6gd6ql3jrc5yzme8v4ntcewwz5cnw92tz0pc8qcuufvq7khhr8wpald05e92xw006sq94mg8v2ndf4sefvf9sygkshp5zfem29trqq2yxxz7
2017-11-08 14:34:11 +01:00
#
2017-10-26 05:05:19 +02:00
# Breakdown:
2017-11-08 14:34:11 +01:00
#
2017-10-26 05:05:19 +02:00
# * `lnbc`: prefix, lightning on bitcoin mainnet
# * `20m`: amount (20 milli-bitcoin)
# * `1`: Bech32 separator
# * `pvjluez`: timestamp (1496314658)
# * `p`: payment hash...
# * `h`: tagged field: hash of description
# * `p5`: `data_length` (`p` = 1, `5` = 20. 1 * 32 + 20 == 52)
# * `8yjmdan79s6qqdhdzgynm4zwqd5d7xmw5fk98klysy043l2ahrqs`: SHA256 of 'One piece of chocolate cake, one icecream cone, one pickle, one slice of swiss cheese, one slice of salami, one lollypop, one piece of cherry pie, one sausage, one cupcake, and one slice of watermelon'
# * `vjfls3ljx9e93jkw0kw40yxn4pevgzflf83qh2852esjddv4xk4z70nehrdcxa4fk0t6hlcc6vrxywke6njenk7yzkzw0quqcwxphkcp`: signature
# * `vam37w`: Bech32 checksum
b11 = l1 . rpc . decodepay ( ' lnbc20m1pvjluezpp5qqqsyqcyq5rqwzqfqqqsyqcyq5rqwzqfqqqsyqcyq5rqwzqfqypqhp58yjmdan79s6qqdhdzgynm4zwqd5d7xmw5fk98klysy043l2ahrqscc6gd6ql3jrc5yzme8v4ntcewwz5cnw92tz0pc8qcuufvq7khhr8wpald05e92xw006sq94mg8v2ndf4sefvf9sygkshp5zfem29trqq2yxxz7 ' , ' One piece of chocolate cake, one icecream cone, one pickle, one slice of swiss cheese, one slice of salami, one lollypop, one piece of cherry pie, one sausage, one cupcake, and one slice of watermelon ' )
assert b11 [ ' currency ' ] == ' bc '
assert b11 [ ' msatoshi ' ] == 20 * 10 * * 11 / / 1000
assert b11 [ ' timestamp ' ] == 1496314658
assert b11 [ ' payment_hash ' ] == ' 0001020304050607080900010203040506070809000102030405060708090102 '
assert b11 [ ' expiry ' ] == 3600
assert b11 [ ' payee ' ] == ' 03e7156ae33b0a208d0744199163177e909e80176e55d97a2f221ede0f934dd9ad '
# > ### The same, on testnet, with a fallback address mk2QpYatsKicvFVuTAQLBryyccRXMUaGHP
# > lntb20m1pvjluezhp58yjmdan79s6qqdhdzgynm4zwqd5d7xmw5fk98klysy043l2ahrqspp5qqqsyqcyq5rqwzqfqqqsyqcyq5rqwzqfqqqsyqcyq5rqwzqfqypqfpp3x9et2e20v6pu37c5d9vax37wxq72un98kmzzhznpurw9sgl2v0nklu2g4d0keph5t7tj9tcqd8rexnd07ux4uv2cjvcqwaxgj7v4uwn5wmypjd5n69z2xm3xgksg28nwht7f6zspwp3f9t
2017-11-08 14:34:11 +01:00
#
2017-10-26 05:05:19 +02:00
# Breakdown:
2017-11-08 14:34:11 +01:00
#
2017-10-26 05:05:19 +02:00
# * `lntb`: prefix, lightning on bitcoin testnet
# * `20m`: amount (20 milli-bitcoin)
# * `1`: Bech32 separator
# * `pvjluez`: timestamp (1496314658)
# * `p`: payment hash...
# * `f`: tagged field: fallback address
# * `pp`: `data_length` (`p` = 1. 1 * 32 + 1 == 33)
# * `3x9et2e20v6pu37c5d9vax37wxq72un98`: `3` = 17, so P2PKH address
# * `h`: tagged field: hash of description...
# * `qh84fmvn2klvglsjxfy0vq2mz6t9kjfzlxfwgljj35w2kwa60qv49k7jlsgx43yhs9nuutllkhhnt090mmenuhp8ue33pv4klmrzlcqp`: signature
# * `us2s2r`: Bech32 checksum
b11 = l1 . rpc . decodepay ( ' lntb20m1pvjluezhp58yjmdan79s6qqdhdzgynm4zwqd5d7xmw5fk98klysy043l2ahrqspp5qqqsyqcyq5rqwzqfqqqsyqcyq5rqwzqfqqqsyqcyq5rqwzqfqypqfpp3x9et2e20v6pu37c5d9vax37wxq72un98kmzzhznpurw9sgl2v0nklu2g4d0keph5t7tj9tcqd8rexnd07ux4uv2cjvcqwaxgj7v4uwn5wmypjd5n69z2xm3xgksg28nwht7f6zspwp3f9t ' , ' One piece of chocolate cake, one icecream cone, one pickle, one slice of swiss cheese, one slice of salami, one lollypop, one piece of cherry pie, one sausage, one cupcake, and one slice of watermelon ' )
assert b11 [ ' currency ' ] == ' tb '
assert b11 [ ' msatoshi ' ] == 20 * 10 * * 11 / / 1000
assert b11 [ ' timestamp ' ] == 1496314658
assert b11 [ ' payment_hash ' ] == ' 0001020304050607080900010203040506070809000102030405060708090102 '
assert b11 [ ' expiry ' ] == 3600
assert b11 [ ' payee ' ] == ' 03e7156ae33b0a208d0744199163177e909e80176e55d97a2f221ede0f934dd9ad '
assert b11 [ ' fallback ' ] [ ' type ' ] == ' P2PKH '
assert b11 [ ' fallback ' ] [ ' addr ' ] == ' mk2QpYatsKicvFVuTAQLBryyccRXMUaGHP '
2017-11-08 14:34:11 +01:00
2017-10-26 05:05:19 +02:00
# > ### On mainnet, with fallback address 1RustyRX2oai4EYYDpQGWvEL62BBGqN9T with extra routing info to go via nodes 029e03a901b85534ff1e92c43c74431f7ce72046060fcf7a95c37e148f78c77255 then 039e03a901b85534ff1e92c43c74431f7ce72046060fcf7a95c37e148f78c77255
# > lnbc20m1pvjluezpp5qqqsyqcyq5rqwzqfqqqsyqcyq5rqwzqfqqqsyqcyq5rqwzqfqypqhp58yjmdan79s6qqdhdzgynm4zwqd5d7xmw5fk98klysy043l2ahrqsfpp3qjmp7lwpagxun9pygexvgpjdc4jdj85fr9yq20q82gphp2nflc7jtzrcazrra7wwgzxqc8u7754cdlpfrmccae92qgzqvzq2ps8pqqqqqqqqqqqq9qqqvpeuqafqxu92d8lr6fvg0r5gv0heeeqgcrqlnm6jhphu9y00rrhy4grqszsvpcgpy9qqqqqqqqqqqq7qqzqfnlkwydm8rg30gjku7wmxmk06sevjp53fmvrcfegvwy7d5443jvyhxsel0hulkstws7vqv400q4j3wgpk4crg49682hr4scqvmad43cqd5m7tf
2017-11-08 14:34:11 +01:00
#
2017-10-26 05:05:19 +02:00
# Breakdown:
2017-11-08 14:34:11 +01:00
#
2017-10-26 05:05:19 +02:00
# * `lnbc`: prefix, lightning on bitcoin mainnet
# * `20m`: amount (20 milli-bitcoin)
# * `1`: Bech32 separator
# * `pvjluez`: timestamp (1496314658)
# * `p`: payment hash...
# * `h`: tagged field: hash of description...
# * `f`: tagged field: fallback address
# * `pp`: `data_length` (`p` = 1. 1 * 32 + 1 == 33)
# * `3qjmp7lwpagxun9pygexvgpjdc4jdj85f`: `3` = 17, so P2PKH address
# * `r`: tagged field: route information
# * `9y`: `data_length` (`9` = 5, `y` = 4. 5 * 32 + 4 = 164)
# `q20q82gphp2nflc7jtzrcazrra7wwgzxqc8u7754cdlpfrmccae92qgzqvzq2ps8pqqqqqqqqqqqq9qqqvpeuqafqxu92d8lr6fvg0r5gv0heeeqgcrqlnm6jhphu9y00rrhy4grqszsvpcgpy9qqqqqqqqqqqq7qqzq`: pubkey `029e03a901b85534ff1e92c43c74431f7ce72046060fcf7a95c37e148f78c77255`, `short_channel_id` 0102030405060708, `fee` 20 millisatoshi, `cltv_expiry_delta` 3. pubkey `039e03a901b85534ff1e92c43c74431f7ce72046060fcf7a95c37e148f78c77255`, `short_channel_id` 030405060708090a, `fee` 30 millisatoshi, `cltv_expiry_delta` 4.
# * `fnlkwydm8rg30gjku7wmxmk06sevjp53fmvrcfegvwy7d5443jvyhxsel0hulkstws7vqv400q4j3wgpk4crg49682hr4scqvmad43cq`: signature
# * `d5m7tf`: Bech32 checksum
b11 = l1 . rpc . decodepay ( ' lnbc20m1pvjluezpp5qqqsyqcyq5rqwzqfqqqsyqcyq5rqwzqfqqqsyqcyq5rqwzqfqypqhp58yjmdan79s6qqdhdzgynm4zwqd5d7xmw5fk98klysy043l2ahrqsfpp3qjmp7lwpagxun9pygexvgpjdc4jdj85fr9yq20q82gphp2nflc7jtzrcazrra7wwgzxqc8u7754cdlpfrmccae92qgzqvzq2ps8pqqqqqqqqqqqq9qqqvpeuqafqxu92d8lr6fvg0r5gv0heeeqgcrqlnm6jhphu9y00rrhy4grqszsvpcgpy9qqqqqqqqqqqq7qqzqfnlkwydm8rg30gjku7wmxmk06sevjp53fmvrcfegvwy7d5443jvyhxsel0hulkstws7vqv400q4j3wgpk4crg49682hr4scqvmad43cqd5m7tf ' , ' One piece of chocolate cake, one icecream cone, one pickle, one slice of swiss cheese, one slice of salami, one lollypop, one piece of cherry pie, one sausage, one cupcake, and one slice of watermelon ' )
assert b11 [ ' currency ' ] == ' bc '
assert b11 [ ' msatoshi ' ] == 20 * 10 * * 11 / / 1000
assert b11 [ ' timestamp ' ] == 1496314658
assert b11 [ ' payment_hash ' ] == ' 0001020304050607080900010203040506070809000102030405060708090102 '
assert b11 [ ' expiry ' ] == 3600
assert b11 [ ' payee ' ] == ' 03e7156ae33b0a208d0744199163177e909e80176e55d97a2f221ede0f934dd9ad '
assert b11 [ ' fallback ' ] [ ' type ' ] == ' P2PKH '
assert b11 [ ' fallback ' ] [ ' addr ' ] == ' 1RustyRX2oai4EYYDpQGWvEL62BBGqN9T '
assert len ( b11 [ ' routes ' ] ) == 1
assert len ( b11 [ ' routes ' ] [ 0 ] ) == 2
assert b11 [ ' routes ' ] [ 0 ] [ 0 ] [ ' pubkey ' ] == ' 029e03a901b85534ff1e92c43c74431f7ce72046060fcf7a95c37e148f78c77255 '
# 0x010203:0x040506:0x0708
assert b11 [ ' routes ' ] [ 0 ] [ 0 ] [ ' short_channel_id ' ] == ' 66051:263430:1800 '
assert b11 [ ' routes ' ] [ 0 ] [ 0 ] [ ' fee ' ] == 20
assert b11 [ ' routes ' ] [ 0 ] [ 0 ] [ ' cltv_expiry_delta ' ] == 3
assert b11 [ ' routes ' ] [ 0 ] [ 1 ] [ ' pubkey ' ] == ' 039e03a901b85534ff1e92c43c74431f7ce72046060fcf7a95c37e148f78c77255 '
# 0x030405:0x060708:0x090a
assert b11 [ ' routes ' ] [ 0 ] [ 1 ] [ ' short_channel_id ' ] == ' 197637:395016:2314 '
assert b11 [ ' routes ' ] [ 0 ] [ 1 ] [ ' fee ' ] == 30
assert b11 [ ' routes ' ] [ 0 ] [ 1 ] [ ' cltv_expiry_delta ' ] == 4
2017-11-08 14:34:11 +01:00
2017-10-26 05:05:19 +02:00
# > ### On mainnet, with fallback (P2SH) address 3EktnHQD7RiAE6uzMj2ZifT9YgRrkSgzQX
# > lnbc20m1pvjluezhp58yjmdan79s6qqdhdzgynm4zwqd5d7xmw5fk98klysy043l2ahrqspp5qqqsyqcyq5rqwzqfqqqsyqcyq5rqwzqfqqqsyqcyq5rqwzqfqypqfppj3a24vwu6r8ejrss3axul8rxldph2q7z9kmrgvr7xlaqm47apw3d48zm203kzcq357a4ls9al2ea73r8jcceyjtya6fu5wzzpe50zrge6ulk4nvjcpxlekvmxl6qcs9j3tz0469gq5g658y
2017-11-08 14:34:11 +01:00
#
2017-10-26 05:05:19 +02:00
# Breakdown:
2017-11-08 14:34:11 +01:00
#
2017-10-26 05:05:19 +02:00
# * `lnbc`: prefix, lightning on bitcoin mainnet
# * `20m`: amount (20 milli-bitcoin)
# * `1`: Bech32 separator
# * `pvjluez`: timestamp (1496314658)
# * `p`: payment hash...
# * `f`: tagged field: fallback address.
# * `pp`: `data_length` (`p` = 1. 1 * 32 + 1 == 33)
# * `j3a24vwu6r8ejrss3axul8rxldph2q7z9`: `j` = 18, so P2SH address
# * `h`: tagged field: hash of description...
# * `2jhz8j78lv2jynuzmz6g8ve53he7pheeype33zlja5azae957585uu7x59w0f2l3rugyva6zpu394y4rh093j6wxze0ldsvk757a9msq`: signature
# * `mf9swh`: Bech32 checksum
b11 = l1 . rpc . decodepay ( ' lnbc20m1pvjluezhp58yjmdan79s6qqdhdzgynm4zwqd5d7xmw5fk98klysy043l2ahrqspp5qqqsyqcyq5rqwzqfqqqsyqcyq5rqwzqfqqqsyqcyq5rqwzqfqypqfppj3a24vwu6r8ejrss3axul8rxldph2q7z9kmrgvr7xlaqm47apw3d48zm203kzcq357a4ls9al2ea73r8jcceyjtya6fu5wzzpe50zrge6ulk4nvjcpxlekvmxl6qcs9j3tz0469gq5g658y ' , ' One piece of chocolate cake, one icecream cone, one pickle, one slice of swiss cheese, one slice of salami, one lollypop, one piece of cherry pie, one sausage, one cupcake, and one slice of watermelon ' )
assert b11 [ ' currency ' ] == ' bc '
assert b11 [ ' msatoshi ' ] == 20 * 10 * * 11 / / 1000
assert b11 [ ' timestamp ' ] == 1496314658
assert b11 [ ' payment_hash ' ] == ' 0001020304050607080900010203040506070809000102030405060708090102 '
assert b11 [ ' expiry ' ] == 3600
assert b11 [ ' payee ' ] == ' 03e7156ae33b0a208d0744199163177e909e80176e55d97a2f221ede0f934dd9ad '
assert b11 [ ' fallback ' ] [ ' type ' ] == ' P2SH '
assert b11 [ ' fallback ' ] [ ' addr ' ] == ' 3EktnHQD7RiAE6uzMj2ZifT9YgRrkSgzQX '
# > ### On mainnet, with fallback (P2WPKH) address bc1qw508d6qejxtdg4y5r3zarvary0c5xw7kv8f3t4
# > lnbc20m1pvjluezhp58yjmdan79s6qqdhdzgynm4zwqd5d7xmw5fk98klysy043l2ahrqspp5qqqsyqcyq5rqwzqfqqqsyqcyq5rqwzqfqqqsyqcyq5rqwzqfqypqfppqw508d6qejxtdg4y5r3zarvary0c5xw7kepvrhrm9s57hejg0p662ur5j5cr03890fa7k2pypgttmh4897d3raaq85a293e9jpuqwl0rnfuwzam7yr8e690nd2ypcq9hlkdwdvycqa0qza8
2017-11-08 14:34:11 +01:00
#
2017-10-26 05:05:19 +02:00
# * `lnbc`: prefix, lightning on bitcoin mainnet
# * `20m`: amount (20 milli-bitcoin)
# * `1`: Bech32 separator
# * `pvjluez`: timestamp (1496314658)
# * `p`: payment hash...
# * `f`: tagged field: fallback address.
# * `pp`: `data_length` (`p` = 1. 1 * 32 + 1 == 33)
2017-11-08 14:34:11 +01:00
# * `q`: 0, so witness version 0.
2017-10-26 05:05:19 +02:00
# * `qw508d6qejxtdg4y5r3zarvary0c5xw7k`: 160 bits = P2WPKH.
# * `h`: tagged field: hash of description...
# * `gw6tk8z0p0qdy9ulggx65lvfsg3nxxhqjxuf2fvmkhl9f4jc74gy44d5ua9us509prqz3e7vjxrftn3jnk7nrglvahxf7arye5llphgq`: signature
# * `qdtpa4`: Bech32 checksum
b11 = l1 . rpc . decodepay ( ' lnbc20m1pvjluezhp58yjmdan79s6qqdhdzgynm4zwqd5d7xmw5fk98klysy043l2ahrqspp5qqqsyqcyq5rqwzqfqqqsyqcyq5rqwzqfqqqsyqcyq5rqwzqfqypqfppqw508d6qejxtdg4y5r3zarvary0c5xw7kepvrhrm9s57hejg0p662ur5j5cr03890fa7k2pypgttmh4897d3raaq85a293e9jpuqwl0rnfuwzam7yr8e690nd2ypcq9hlkdwdvycqa0qza8 ' , ' One piece of chocolate cake, one icecream cone, one pickle, one slice of swiss cheese, one slice of salami, one lollypop, one piece of cherry pie, one sausage, one cupcake, and one slice of watermelon ' )
assert b11 [ ' currency ' ] == ' bc '
assert b11 [ ' msatoshi ' ] == 20 * 10 * * 11 / / 1000
assert b11 [ ' timestamp ' ] == 1496314658
assert b11 [ ' payment_hash ' ] == ' 0001020304050607080900010203040506070809000102030405060708090102 '
assert b11 [ ' expiry ' ] == 3600
assert b11 [ ' payee ' ] == ' 03e7156ae33b0a208d0744199163177e909e80176e55d97a2f221ede0f934dd9ad '
assert b11 [ ' fallback ' ] [ ' type ' ] == ' P2WPKH '
assert b11 [ ' fallback ' ] [ ' addr ' ] == ' bc1qw508d6qejxtdg4y5r3zarvary0c5xw7kv8f3t4 '
# > ### On mainnet, with fallback (P2WSH) address bc1qrp33g0q5c5txsp9arysrx4k6zdkfs4nce4xj0gdcccefvpysxf3qccfmv3
# > lnbc20m1pvjluezhp58yjmdan79s6qqdhdzgynm4zwqd5d7xmw5fk98klysy043l2ahrqspp5qqqsyqcyq5rqwzqfqqqsyqcyq5rqwzqfqqqsyqcyq5rqwzqfqypqfp4qrp33g0q5c5txsp9arysrx4k6zdkfs4nce4xj0gdcccefvpysxf3q28j0v3rwgy9pvjnd48ee2pl8xrpxysd5g44td63g6xcjcu003j3qe8878hluqlvl3km8rm92f5stamd3jw763n3hck0ct7p8wwj463cql26ava
2017-11-08 14:34:11 +01:00
#
2017-10-26 05:05:19 +02:00
# * `lnbc`: prefix, lightning on bitcoin mainnet
# * `20m`: amount (20 milli-bitcoin)
# * `1`: Bech32 separator
# * `pvjluez`: timestamp (1496314658)
# * `p`: payment hash...
# * `f`: tagged field: fallback address.
# * `p4`: `data_length` (`p` = 1, `4` = 21. 1 * 32 + 21 == 53)
# * `q`: 0, so witness version 0.
# * `rp33g0q5c5txsp9arysrx4k6zdkfs4nce4xj0gdcccefvpysxf3q`: 260 bits = P2WSH.
# * `h`: tagged field: hash of description...
# * `5yps56lmsvgcrf476flet6js02m93kgasews8q3jhtp7d6cqckmh70650maq4u65tk53ypszy77v9ng9h2z3q3eqhtc3ewgmmv2grasp`: signature
# * `akvd7y`: Bech32 checksum
b11 = l1 . rpc . decodepay ( ' lnbc20m1pvjluezhp58yjmdan79s6qqdhdzgynm4zwqd5d7xmw5fk98klysy043l2ahrqspp5qqqsyqcyq5rqwzqfqqqsyqcyq5rqwzqfqqqsyqcyq5rqwzqfqypqfp4qrp33g0q5c5txsp9arysrx4k6zdkfs4nce4xj0gdcccefvpysxf3q28j0v3rwgy9pvjnd48ee2pl8xrpxysd5g44td63g6xcjcu003j3qe8878hluqlvl3km8rm92f5stamd3jw763n3hck0ct7p8wwj463cql26ava ' , ' One piece of chocolate cake, one icecream cone, one pickle, one slice of swiss cheese, one slice of salami, one lollypop, one piece of cherry pie, one sausage, one cupcake, and one slice of watermelon ' )
assert b11 [ ' currency ' ] == ' bc '
assert b11 [ ' msatoshi ' ] == 20 * 10 * * 11 / / 1000
assert b11 [ ' timestamp ' ] == 1496314658
assert b11 [ ' payment_hash ' ] == ' 0001020304050607080900010203040506070809000102030405060708090102 '
assert b11 [ ' expiry ' ] == 3600
assert b11 [ ' payee ' ] == ' 03e7156ae33b0a208d0744199163177e909e80176e55d97a2f221ede0f934dd9ad '
assert b11 [ ' fallback ' ] [ ' type ' ] == ' P2WSH '
assert b11 [ ' fallback ' ] [ ' addr ' ] == ' bc1qrp33g0q5c5txsp9arysrx4k6zdkfs4nce4xj0gdcccefvpysxf3qccfmv3 '
2017-04-29 03:02:27 +02:00
def test_sendpay ( self ) :
l1 , l2 = self . connect ( )
self . fund_channel ( l1 , l2 , 10 * * 6 )
amt = 200000000
2017-10-26 05:06:19 +02:00
rhash = l2 . rpc . invoice ( amt , ' testpayment2 ' , ' desc ' ) [ ' rhash ' ]
2017-04-29 03:02:27 +02:00
assert l2 . rpc . listinvoice ( ' testpayment2 ' ) [ 0 ] [ ' complete ' ] == False
2017-04-29 13:28:45 +02:00
routestep = {
' msatoshi ' : amt ,
' id ' : l2 . info [ ' id ' ] ,
' delay ' : 5 ,
' channel ' : ' 1:1:1 '
}
2017-04-29 03:02:27 +02:00
# Insufficient funds.
rs = copy . deepcopy ( routestep )
rs [ ' msatoshi ' ] = rs [ ' msatoshi ' ] - 1
self . assertRaises ( ValueError , l1 . rpc . sendpay , to_json ( [ rs ] ) , rhash )
assert l2 . rpc . listinvoice ( ' testpayment2 ' ) [ 0 ] [ ' complete ' ] == False
# Gross overpayment (more than factor of 2)
rs = copy . deepcopy ( routestep )
rs [ ' msatoshi ' ] = rs [ ' msatoshi ' ] * 2 + 1
self . assertRaises ( ValueError , l1 . rpc . sendpay , to_json ( [ rs ] ) , rhash )
assert l2 . rpc . listinvoice ( ' testpayment2 ' ) [ 0 ] [ ' complete ' ] == False
2017-08-14 22:36:42 +02:00
2017-04-29 03:02:27 +02:00
# Insufficient delay.
rs = copy . deepcopy ( routestep )
rs [ ' delay ' ] = rs [ ' delay ' ] - 2
self . assertRaises ( ValueError , l1 . rpc . sendpay , to_json ( [ rs ] ) , rhash )
assert l2 . rpc . listinvoice ( ' testpayment2 ' ) [ 0 ] [ ' complete ' ] == False
# Bad ID.
rs = copy . deepcopy ( routestep )
rs [ ' id ' ] = ' 00000000000000000000000000000000 '
self . assertRaises ( ValueError , l1 . rpc . sendpay , to_json ( [ rs ] ) , rhash )
assert l2 . rpc . listinvoice ( ' testpayment2 ' ) [ 0 ] [ ' complete ' ] == False
2017-06-20 07:59:03 +02:00
# FIXME: test paying via another node, should fail to pay twice.
p1 = l1 . rpc . getpeer ( l2 . info [ ' id ' ] , ' info ' )
p2 = l2 . rpc . getpeer ( l1 . info [ ' id ' ] , ' info ' )
assert p1 [ ' msatoshi_to_us ' ] == 10 * * 6 * 1000
assert p1 [ ' msatoshi_total ' ] == 10 * * 6 * 1000
assert p2 [ ' msatoshi_to_us ' ] == 0
assert p2 [ ' msatoshi_total ' ] == 10 * * 6 * 1000
2017-04-29 03:02:27 +02:00
# This works.
l1 . rpc . sendpay ( to_json ( [ routestep ] ) , rhash )
assert l2 . rpc . listinvoice ( ' testpayment2 ' ) [ 0 ] [ ' complete ' ] == True
2017-06-20 07:59:03 +02:00
# Balances should reflect it.
2017-08-14 22:36:42 +02:00
time . sleep ( 1 )
2017-06-20 07:59:03 +02:00
p1 = l1 . rpc . getpeer ( l2 . info [ ' id ' ] , ' info ' )
p2 = l2 . rpc . getpeer ( l1 . info [ ' id ' ] , ' info ' )
assert p1 [ ' msatoshi_to_us ' ] == 10 * * 6 * 1000 - amt
assert p1 [ ' msatoshi_total ' ] == 10 * * 6 * 1000
assert p2 [ ' msatoshi_to_us ' ] == amt
assert p2 [ ' msatoshi_total ' ] == 10 * * 6 * 1000
2017-04-29 03:02:27 +02:00
# Repeat will "succeed", but won't actually send anything (duplicate)
assert not l1 . daemon . is_in_log ( ' ... succeeded ' )
l1 . rpc . sendpay ( to_json ( [ routestep ] ) , rhash )
l1 . daemon . wait_for_log ( ' ... succeeded ' )
assert l2 . rpc . listinvoice ( ' testpayment2 ' ) [ 0 ] [ ' complete ' ] == True
# Overpaying by "only" a factor of 2 succeeds.
2017-10-26 05:06:19 +02:00
rhash = l2 . rpc . invoice ( amt , ' testpayment3 ' , ' desc ' ) [ ' rhash ' ]
2017-04-29 03:02:27 +02:00
assert l2 . rpc . listinvoice ( ' testpayment3 ' ) [ 0 ] [ ' complete ' ] == False
2017-04-29 13:28:45 +02:00
routestep = { ' msatoshi ' : amt * 2 , ' id ' : l2 . info [ ' id ' ] , ' delay ' : 5 , ' channel ' : ' 1:1:1 ' }
2017-04-29 03:02:27 +02:00
l1 . rpc . sendpay ( to_json ( [ routestep ] ) , rhash )
assert l2 . rpc . listinvoice ( ' testpayment3 ' ) [ 0 ] [ ' complete ' ] == True
2017-08-30 03:21:48 +02:00
def test_sendpay_cant_afford ( self ) :
l1 , l2 = self . connect ( )
# Note, this is in SATOSHI, rest are in MILLISATOSHI!
self . fund_channel ( l1 , l2 , 10 * * 6 )
# Can't pay more than channel capacity.
self . assertRaises ( ValueError , self . pay , l1 , l2 , 10 * * 9 + 1 )
# This is the fee, which needs to be taken into account for l1.
available = 10 * * 9 - 13440
# Reserve is 1%.
reserve = 10 * * 7
# Can't pay past reserve.
self . assertRaises ( ValueError , self . pay , l1 , l2 , available )
self . assertRaises ( ValueError , self . pay , l1 , l2 , available - reserve + 1 )
# Can pay up to reserve (1%)
self . pay ( l1 , l2 , available - reserve )
# And now it can't pay back, due to its own reserve.
self . assertRaises ( ValueError , self . pay , l2 , l1 , available - reserve )
# But this should work.
self . pay ( l2 , l1 , available - reserve * 2 )
2017-10-26 05:07:19 +02:00
def test_pay ( self ) :
l1 , l2 = self . connect ( )
chanid = self . fund_channel ( l1 , l2 , 10 * * 6 )
# Wait for route propagation.
2017-11-28 05:39:22 +01:00
bitcoind . generate_block ( 5 )
2017-10-26 05:07:19 +02:00
l1 . daemon . wait_for_logs ( [ ' Received channel_update for channel {} \ (0 \ ) '
. format ( chanid ) ,
' Received channel_update for channel {} \ (1 \ ) '
. format ( chanid ) ] )
inv = l2 . rpc . invoice ( 123000 , ' test_pay ' , ' description ' ) [ ' bolt11 ' ]
l1 . rpc . pay ( inv ) ;
assert l2 . rpc . listinvoice ( ' test_pay ' ) [ 0 ] [ ' complete ' ] == True
2017-09-28 05:53:23 +02:00
def test_bad_opening ( self ) :
# l1 asks for a too-long locktime
2017-10-09 03:30:00 +02:00
l1 = self . node_factory . get_node ( options = [ ' --locktime-blocks=100 ' ] )
l2 = self . node_factory . get_node ( options = [ ' --max-locktime-blocks=99 ' ] )
2017-12-04 10:12:22 +01:00
ret = l1 . rpc . connect ( l2 . info [ ' id ' ] , ' localhost ' , l2 . info [ ' port ' ] )
2017-09-28 05:53:23 +02:00
assert ret [ ' id ' ] == l2 . info [ ' id ' ]
gossipd: rewrite to do the handshake internally.
Now the flow is much simpler from a lightningd POV:
1. If we want to connect to a peer, just send gossipd `gossipctl_reach_peer`.
2. Every new peer, gossipd hands up to lightningd, with global/local features
and the peer fd and a gossip fd using `gossip_peer_connected`
3. If lightningd doesn't want it, it just hands the peerfd and global/local
features back to gossipd using `gossipctl_handle_peer`
4. If a peer sends a non-gossip msg (eg `open_channel`) the gossipd sends
it up using `gossip_peer_nongossip`.
5. If lightningd wants to fund a channel, it simply calls `release_channel`.
Notes:
* There's no more "unique_id": we use the peer id.
* For the moment, we don't ask gossipd when we're told to list peers, so
connected peers without a channel don't appear in the JSON getpeers API.
* We add a `gossipctl_peer_addrhint` for the moment, so you can connect to
a specific ip/port, but using other sources is a TODO.
* We now (correctly) only give up on reaching a peer after we exchange init
messages, which changes the test_disconnect case.
Signed-off-by: Rusty Russell <rusty@rustcorp.com.au>
2017-10-11 12:09:49 +02:00
l1 . daemon . wait_for_log ( ' WIRE_GOSSIPCTL_HANDLE_PEER ' )
l2 . daemon . wait_for_log ( ' WIRE_GOSSIPCTL_HANDLE_PEER ' )
2017-09-28 05:53:23 +02:00
addr = l1 . rpc . newaddr ( ) [ ' address ' ]
txid = l1 . bitcoin . rpc . sendtoaddress ( addr , 10 * * 6 / 10 * * 8 + 0.01 )
tx = l1 . bitcoin . rpc . getrawtransaction ( txid )
l1 . rpc . addfunds ( tx )
self . assertRaises ( ValueError , l1 . rpc . fundchannel , l2 . info [ ' id ' ] , 10 * * 6 )
l2 . daemon . wait_for_log ( ' to_self_delay 100 larger than 99 ' )
2017-06-26 03:16:43 +02:00
def test_closing ( self ) :
l1 , l2 = self . connect ( )
self . fund_channel ( l1 , l2 , 10 * * 6 )
2017-07-05 07:33:40 +02:00
self . pay ( l1 , l2 , 200000000 )
2017-06-26 03:16:43 +02:00
2017-07-04 02:48:43 +02:00
assert l1 . bitcoin . rpc . getmempoolinfo ( ) [ ' size ' ] == 0
2017-06-26 03:16:43 +02:00
# This should return, then close.
l1 . rpc . close ( l2 . info [ ' id ' ] ) ;
l1 . daemon . wait_for_log ( ' -> CHANNELD_SHUTTING_DOWN ' )
l2 . daemon . wait_for_log ( ' -> CHANNELD_SHUTTING_DOWN ' )
2017-04-29 03:02:27 +02:00
2017-07-04 02:47:32 +02:00
l1 . daemon . wait_for_log ( ' -> CLOSINGD_SIGEXCHANGE ' )
l2 . daemon . wait_for_log ( ' -> CLOSINGD_SIGEXCHANGE ' )
2017-07-04 03:53:35 +02:00
# And should put closing into mempool.
l1 . daemon . wait_for_log ( ' sendrawtx exit 0 ' )
l2 . daemon . wait_for_log ( ' sendrawtx exit 0 ' )
assert l1 . bitcoin . rpc . getmempoolinfo ( ) [ ' size ' ] == 1
2017-08-14 22:36:42 +02:00
2017-10-24 04:06:14 +02:00
@unittest.skipIf ( not DEVELOPER , " needs DEVELOPER=1 " )
2017-08-18 06:43:52 +02:00
def test_permfail ( self ) :
l1 , l2 = self . connect ( )
self . fund_channel ( l1 , l2 , 10 * * 6 )
self . pay ( l1 , l2 , 200000000 )
2017-08-23 03:55:23 +02:00
# Make sure l2 has received sig with 0 htlcs!
l2 . daemon . wait_for_log ( ' Received commit_sig with 0 htlc sigs ' )
# Make sure l1 has final revocation.
l1 . daemon . wait_for_log ( ' Sending commit_sig with 0 htlc sigs ' )
l1 . daemon . wait_for_log ( ' peer_in WIRE_REVOKE_AND_ACK ' )
2017-08-18 06:43:52 +02:00
# We fail l2, so l1 will reconnect to it.
l2 . rpc . dev_fail ( l1 . info [ ' id ' ] ) ;
l2 . daemon . wait_for_log ( ' Failing due to dev-fail command ' )
l2 . daemon . wait_for_log ( ' sendrawtx exit 0 ' )
# "Internal error" in hex
l1 . daemon . wait_for_log ( ' WIRE_ERROR.*496e7465726e616c206572726f72 ' )
2017-08-23 03:55:23 +02:00
# l2 will send out tx (l1 considers it a transient error)
2017-11-28 05:39:22 +01:00
bitcoind . generate_block ( 1 )
2017-08-23 03:55:23 +02:00
l1 . daemon . wait_for_log ( ' Their unilateral tx, old commit point ' )
l1 . daemon . wait_for_log ( ' -> ONCHAIND_THEIR_UNILATERAL ' )
l2 . daemon . wait_for_log ( ' -> ONCHAIND_OUR_UNILATERAL ' )
2017-10-23 06:16:57 +02:00
l2 . daemon . wait_for_log ( ' Propose handling OUR_UNILATERAL/DELAYED_OUTPUT_TO_US by OUR_DELAYED_RETURN_TO_WALLET (.*) in 5 blocks ' )
2017-08-23 03:55:23 +02:00
2017-10-23 06:16:57 +02:00
# Now, mine 5 blocks so it sends out the spending tx.
2017-11-28 05:39:22 +01:00
bitcoind . generate_block ( 5 )
2017-08-23 03:55:23 +02:00
# It should send the to-wallet tx.
2017-09-20 06:54:16 +02:00
l2 . daemon . wait_for_log ( ' Broadcasting OUR_DELAYED_RETURN_TO_WALLET ' )
2017-08-23 03:55:23 +02:00
l2 . daemon . wait_for_log ( ' sendrawtx exit 0 ' )
# 100 after l1 sees tx, it should be done.
2017-11-28 05:39:22 +01:00
bitcoind . generate_block ( 95 )
2017-08-23 03:55:23 +02:00
l1 . daemon . wait_for_log ( ' onchaind complete, forgetting peer ' )
2017-09-26 23:02:37 +02:00
# Now, 100 blocks l2 should be done.
2017-11-28 05:39:22 +01:00
bitcoind . generate_block ( 5 )
2017-08-23 03:55:23 +02:00
l2 . daemon . wait_for_log ( ' onchaind complete, forgetting peer ' )
2017-10-24 04:06:14 +02:00
@unittest.skipIf ( not DEVELOPER , " needs DEVELOPER=1 " )
2017-09-26 23:02:48 +02:00
def test_onchain_first_commit ( self ) :
""" Onchain handling where funder immediately drops to chain """
# HTLC 1->2, 1 fails just after funding.
disconnects = [ ' +WIRE_FUNDING_LOCKED ' , ' permfail ' ]
l1 = self . node_factory . get_node ( disconnect = disconnects )
2017-12-04 03:09:04 +01:00
# Make locktime different, as we once had them reversed!
l2 = self . node_factory . get_node ( options = [ ' --locktime-blocks=10 ' ] )
2017-09-26 23:02:48 +02:00
2017-12-04 10:12:22 +01:00
l1 . rpc . connect ( l2 . info [ ' id ' ] , ' localhost ' , l2 . info [ ' port ' ] )
2017-09-26 23:02:48 +02:00
# Like fundchannel, but we'll probably fail before CHANNELD_NORMAL.
addr = l1 . rpc . newaddr ( ) [ ' address ' ]
txid = l1 . bitcoin . rpc . sendtoaddress ( addr , 10 * * 6 / 10 * * 8 + 0.01 )
tx = l1 . bitcoin . rpc . getrawtransaction ( txid )
l1 . rpc . addfunds ( tx )
l1 . rpc . fundchannel ( l2 . info [ ' id ' ] , 10 * * 6 )
l1 . daemon . wait_for_log ( ' sendrawtx exit 0 ' )
2017-11-28 05:39:22 +01:00
l1 . bitcoin . generate_block ( 1 )
2017-09-26 23:02:48 +02:00
# l1 will drop to chain.
l1 . daemon . wait_for_log ( ' permfail ' )
l1 . daemon . wait_for_log ( ' sendrawtx exit 0 ' )
2017-11-28 05:39:22 +01:00
l1 . bitcoin . generate_block ( 1 )
2017-09-26 23:02:48 +02:00
l1 . daemon . wait_for_log ( ' -> ONCHAIND_OUR_UNILATERAL ' )
l2 . daemon . wait_for_log ( ' -> ONCHAIND_THEIR_UNILATERAL ' )
2017-12-04 03:09:04 +01:00
# 10 later, l1 should collect its to-self payment.
bitcoind . generate_block ( 10 )
2017-09-26 23:02:48 +02:00
l1 . daemon . wait_for_log ( ' Broadcasting OUR_DELAYED_RETURN_TO_WALLET .* to resolve OUR_UNILATERAL/DELAYED_OUTPUT_TO_US ' )
l1 . daemon . wait_for_log ( ' sendrawtx exit 0 ' )
# 94 later, l2 is done.
2017-11-28 05:39:22 +01:00
bitcoind . generate_block ( 94 )
2017-09-26 23:02:48 +02:00
l2 . daemon . wait_for_log ( ' onchaind complete, forgetting peer ' )
# Now, 100 blocks and l1 should be done.
2017-11-28 05:39:22 +01:00
bitcoind . generate_block ( 6 )
2017-09-26 23:02:48 +02:00
l1 . daemon . wait_for_log ( ' onchaind complete, forgetting peer ' )
2017-10-24 04:06:14 +02:00
@unittest.skipIf ( not DEVELOPER , " needs DEVELOPER=1 " )
2017-09-26 23:02:47 +02:00
def test_onchain_dust_out ( self ) :
""" Onchain handling of outgoing dust htlcs (they should fail) """
# HTLC 1->2, 1 fails after it's irrevocably committed
disconnects = [ ' @WIRE_REVOKE_AND_ACK ' , ' permfail ' ]
l1 = self . node_factory . get_node ( disconnect = disconnects )
l2 = self . node_factory . get_node ( )
2017-12-04 10:12:22 +01:00
l1 . rpc . connect ( l2 . info [ ' id ' ] , ' localhost ' , l2 . info [ ' port ' ] )
2017-09-26 23:02:47 +02:00
self . fund_channel ( l1 , l2 , 10 * * 6 )
# Must be dust!
2017-10-26 05:06:19 +02:00
rhash = l2 . rpc . invoice ( 1 , ' onchain_dust_out ' , ' desc ' ) [ ' rhash ' ]
2017-09-26 23:02:47 +02:00
routestep = {
' msatoshi ' : 1 ,
' id ' : l2 . info [ ' id ' ] ,
' delay ' : 5 ,
' channel ' : ' 1:1:1 '
}
2017-09-28 18:09:38 +02:00
payfuture = self . executor . submit ( l1 . rpc . sendpay , to_json ( [ routestep ] ) , rhash )
2017-09-26 23:02:47 +02:00
# l1 will drop to chain.
l1 . daemon . wait_for_log ( ' permfail ' )
l1 . daemon . wait_for_log ( ' sendrawtx exit 0 ' )
2017-11-28 05:39:22 +01:00
l1 . bitcoin . generate_block ( 1 )
2017-09-26 23:02:47 +02:00
l1 . daemon . wait_for_log ( ' -> ONCHAIND_OUR_UNILATERAL ' )
l2 . daemon . wait_for_log ( ' -> ONCHAIND_THEIR_UNILATERAL ' )
# We use 3 blocks for "reasonable depth"
2017-11-28 05:39:22 +01:00
bitcoind . generate_block ( 3 )
2017-09-26 23:02:47 +02:00
# It should fail.
2017-09-28 18:09:38 +02:00
self . assertRaises ( ValueError , payfuture . result , 5 )
2017-09-26 23:02:47 +02:00
l1 . daemon . wait_for_log ( ' WIRE_PERMANENT_CHANNEL_FAILURE: missing in commitment tx ' )
# 6 later, l1 should collect its to-self payment.
2017-11-28 05:39:22 +01:00
bitcoind . generate_block ( 6 )
2017-09-26 23:02:47 +02:00
l1 . daemon . wait_for_log ( ' Broadcasting OUR_DELAYED_RETURN_TO_WALLET .* to resolve OUR_UNILATERAL/DELAYED_OUTPUT_TO_US ' )
l1 . daemon . wait_for_log ( ' sendrawtx exit 0 ' )
# 94 later, l2 is done.
2017-11-28 05:39:22 +01:00
bitcoind . generate_block ( 94 )
2017-09-26 23:02:47 +02:00
l2 . daemon . wait_for_log ( ' onchaind complete, forgetting peer ' )
# Now, 100 blocks and l1 should be done.
2017-11-28 05:39:22 +01:00
bitcoind . generate_block ( 6 )
2017-09-26 23:02:47 +02:00
l1 . daemon . wait_for_log ( ' onchaind complete, forgetting peer ' )
# Payment failed, BTW
assert l2 . rpc . listinvoice ( ' onchain_dust_out ' ) [ 0 ] [ ' complete ' ] == False
2017-10-24 04:06:14 +02:00
@unittest.skipIf ( not DEVELOPER , " needs DEVELOPER=1 " )
2017-09-26 23:02:47 +02:00
def test_onchain_timeout ( self ) :
""" Onchain handling of outgoing failed htlcs """
# HTLC 1->2, 1 fails just after it's irrevocably committed
disconnects = [ ' +WIRE_REVOKE_AND_ACK ' , ' permfail ' ]
l1 = self . node_factory . get_node ( disconnect = disconnects )
l2 = self . node_factory . get_node ( )
2017-12-04 10:12:22 +01:00
l1 . rpc . connect ( l2 . info [ ' id ' ] , ' localhost ' , l2 . info [ ' port ' ] )
2017-09-26 23:02:47 +02:00
self . fund_channel ( l1 , l2 , 10 * * 6 )
2017-10-26 05:06:19 +02:00
rhash = l2 . rpc . invoice ( 10 * * 8 , ' onchain_timeout ' , ' desc ' ) [ ' rhash ' ]
2017-09-26 23:02:47 +02:00
# We underpay, so it fails.
routestep = {
' msatoshi ' : 10 * * 8 - 1 ,
' id ' : l2 . info [ ' id ' ] ,
' delay ' : 5 ,
' channel ' : ' 1:1:1 '
}
2017-09-28 18:09:38 +02:00
payfuture = self . executor . submit ( l1 . rpc . sendpay , to_json ( [ routestep ] ) , rhash )
2017-09-26 23:02:47 +02:00
# l1 will drop to chain.
l1 . daemon . wait_for_log ( ' permfail ' )
l1 . daemon . wait_for_log ( ' sendrawtx exit 0 ' )
2017-11-28 05:39:22 +01:00
l1 . bitcoin . generate_block ( 1 )
2017-09-26 23:02:47 +02:00
l1 . daemon . wait_for_log ( ' -> ONCHAIND_OUR_UNILATERAL ' )
l2 . daemon . wait_for_log ( ' -> ONCHAIND_THEIR_UNILATERAL ' )
# Wait for timeout.
2017-10-23 06:16:57 +02:00
l1 . daemon . wait_for_log ( ' Propose handling OUR_UNILATERAL/DELAYED_OUTPUT_TO_US by OUR_DELAYED_RETURN_TO_WALLET .* in 5 blocks ' )
2017-11-28 05:39:22 +01:00
bitcoind . generate_block ( 5 )
2017-09-26 23:02:47 +02:00
# (l1 will also collect its to-self payment.)
l1 . daemon . wait_for_log ( ' sendrawtx exit 0 ' )
l1 . daemon . wait_for_log ( ' sendrawtx exit 0 ' )
2017-09-30 22:41:53 +02:00
2017-09-26 23:02:47 +02:00
# We use 3 blocks for "reasonable depth"
2017-11-28 05:39:22 +01:00
bitcoind . generate_block ( 3 )
2017-09-26 23:02:47 +02:00
# It should fail.
2017-09-28 18:09:38 +02:00
self . assertRaises ( ValueError , payfuture . result , 5 )
2017-09-26 23:02:47 +02:00
l1 . daemon . wait_for_log ( ' WIRE_PERMANENT_CHANNEL_FAILURE: timed out ' )
# 91 later, l2 is done.
2017-11-28 05:39:22 +01:00
bitcoind . generate_block ( 91 )
2017-09-26 23:02:47 +02:00
l2 . daemon . wait_for_log ( ' onchaind complete, forgetting peer ' )
# Now, 100 blocks and l1 should be done.
2017-11-28 05:39:22 +01:00
bitcoind . generate_block ( 6 )
2017-09-26 23:02:47 +02:00
l1 . daemon . wait_for_log ( ' onchaind complete, forgetting peer ' )
# Payment failed, BTW
assert l2 . rpc . listinvoice ( ' onchain_timeout ' ) [ 0 ] [ ' complete ' ] == False
2017-10-24 04:06:14 +02:00
@unittest.skipIf ( not DEVELOPER , " needs DEVELOPER=1 " )
2017-09-20 06:45:41 +02:00
def test_onchain_middleman ( self ) :
# HTLC 1->2->3, 1->2 goes down after 2 gets preimage from 3.
disconnects = [ ' -WIRE_UPDATE_FULFILL_HTLC ' , ' permfail ' ]
l1 = self . node_factory . get_node ( )
l2 = self . node_factory . get_node ( disconnect = disconnects )
l3 = self . node_factory . get_node ( )
# l2 connects to both, so l1 can't reconnect and thus l2 drops to chain
2017-12-04 10:12:22 +01:00
l2 . rpc . connect ( l1 . info [ ' id ' ] , ' localhost ' , l1 . info [ ' port ' ] )
l2 . rpc . connect ( l3 . info [ ' id ' ] , ' localhost ' , l3 . info [ ' port ' ] )
2017-09-20 06:45:41 +02:00
self . fund_channel ( l2 , l1 , 10 * * 6 )
self . fund_channel ( l2 , l3 , 10 * * 6 )
# Give l1 some money to play with.
self . pay ( l2 , l1 , 2 * 10 * * 8 )
# Must be bigger than dust!
2017-10-26 05:06:19 +02:00
rhash = l3 . rpc . invoice ( 10 * * 8 , ' middleman ' , ' desc ' ) [ ' rhash ' ]
2017-09-20 06:45:41 +02:00
# Wait for route propagation.
2017-11-28 05:39:22 +01:00
l1 . bitcoin . generate_block ( 5 )
2017-09-20 06:45:41 +02:00
l1 . daemon . wait_for_log ( ' Received node_announcement for node {} '
. format ( l3 . info [ ' id ' ] ) )
route = l1 . rpc . getroute ( l3 . info [ ' id ' ] , 10 * * 8 , 1 ) [ " route " ]
assert len ( route ) == 2
2017-09-26 06:57:19 +02:00
q = queue . Queue ( )
2017-09-20 06:45:41 +02:00
def try_pay ( ) :
2017-09-26 06:57:19 +02:00
try :
l1 . rpc . sendpay ( to_json ( route ) , rhash , async = False )
q . put ( None )
except Exception as err :
q . put ( err )
2017-09-20 06:45:41 +02:00
t = threading . Thread ( target = try_pay )
t . daemon = True
t . start ( )
# l2 will drop to chain.
l2 . daemon . wait_for_log ( ' sendrawtx exit 0 ' )
2017-11-28 05:39:22 +01:00
l1 . bitcoin . generate_block ( 1 )
2017-09-20 06:45:41 +02:00
l2 . daemon . wait_for_log ( ' -> ONCHAIND_OUR_UNILATERAL ' )
l1 . daemon . wait_for_log ( ' -> ONCHAIND_THEIR_UNILATERAL ' )
l2 . daemon . wait_for_log ( ' OUR_UNILATERAL/THEIR_HTLC ' )
2017-09-20 06:54:16 +02:00
# l2 should fulfill HTLC onchain, and spend to-us (any order)
2017-09-20 06:45:41 +02:00
l2 . daemon . wait_for_log ( ' Propose handling OUR_UNILATERAL/THEIR_HTLC by OUR_HTLC_SUCCESS_TX .* in 0 blocks ' )
l2 . daemon . wait_for_log ( ' sendrawtx exit 0 ' )
2017-09-26 06:57:19 +02:00
# Payment should succeed.
2017-11-28 05:39:22 +01:00
l1 . bitcoin . generate_block ( 1 )
2017-09-26 06:57:19 +02:00
l1 . daemon . wait_for_log ( ' THEIR_UNILATERAL/OUR_HTLC gave us preimage ' )
err = q . get ( timeout = 10 )
if err :
print ( " Got err from sendpay thread " )
raise err
t . join ( timeout = 1 )
assert not t . isAlive ( )
2017-10-23 06:16:57 +02:00
# Three more, l2 can spend to-us.
2017-11-28 05:39:22 +01:00
bitcoind . generate_block ( 3 )
2017-09-20 06:54:16 +02:00
l2 . daemon . wait_for_log ( ' Broadcasting OUR_DELAYED_RETURN_TO_WALLET .* to resolve OUR_UNILATERAL/DELAYED_OUTPUT_TO_US ' )
l2 . daemon . wait_for_log ( ' sendrawtx exit 0 ' )
2017-10-23 06:16:57 +02:00
# One more block, HTLC tx is now spentable.
2017-11-28 05:39:22 +01:00
l1 . bitcoin . generate_block ( 1 )
2017-09-20 06:54:16 +02:00
l2 . daemon . wait_for_log ( ' Broadcasting OUR_DELAYED_RETURN_TO_WALLET .* to resolve OUR_HTLC_SUCCESS_TX/DELAYED_OUTPUT_TO_US ' )
2017-09-20 06:45:41 +02:00
l2 . daemon . wait_for_log ( ' sendrawtx exit 0 ' )
# 100 blocks after last spend, l2 should be done.
2017-11-28 05:39:22 +01:00
l1 . bitcoin . generate_block ( 100 )
2017-09-20 06:45:41 +02:00
l2 . daemon . wait_for_log ( ' onchaind complete, forgetting peer ' )
2017-10-24 04:06:14 +02:00
@unittest.skipIf ( not DEVELOPER , " needs DEVELOPER=1 " )
2017-09-26 23:02:37 +02:00
def test_penalty_inhtlc ( self ) :
""" Test penalty transaction with an incoming HTLC """
# We suppress each one after first commit; HTLC gets added not fulfilled.
2017-10-11 12:08:50 +02:00
l1 = self . node_factory . get_node ( disconnect = [ ' =WIRE_COMMITMENT_SIGNED-nocommit ' ] , may_fail = True )
l2 = self . node_factory . get_node ( disconnect = [ ' =WIRE_COMMITMENT_SIGNED-nocommit ' ] )
2017-09-26 23:02:37 +02:00
2017-12-04 10:12:22 +01:00
l1 . rpc . connect ( l2 . info [ ' id ' ] , ' localhost ' , l2 . info [ ' port ' ] )
2017-09-26 23:02:37 +02:00
self . fund_channel ( l1 , l2 , 10 * * 6 )
# Now, this will get stuck due to l1 commit being disabled..
t = self . pay ( l1 , l2 , 100000000 , async = True )
# They should both have commitments blocked now.
2017-10-11 12:08:50 +02:00
l1 . daemon . wait_for_log ( ' =WIRE_COMMITMENT_SIGNED-nocommit ' )
l2 . daemon . wait_for_log ( ' =WIRE_COMMITMENT_SIGNED-nocommit ' )
2017-09-26 23:02:37 +02:00
# Make sure l1 got l2's commitment to the HTLC, and sent to master.
l1 . daemon . wait_for_log ( ' UPDATE WIRE_CHANNEL_GOT_COMMITSIG ' )
2017-09-30 22:41:53 +02:00
2017-09-26 23:02:37 +02:00
# Take our snapshot.
tx = l1 . rpc . dev_sign_last_tx ( l2 . info [ ' id ' ] ) [ ' tx ' ]
# Let them continue
l1 . rpc . dev_reenable_commit ( l2 . info [ ' id ' ] )
l2 . rpc . dev_reenable_commit ( l1 . info [ ' id ' ] )
# Should fulfill.
l1 . daemon . wait_for_log ( ' peer_in WIRE_UPDATE_FULFILL_HTLC ' )
l1 . daemon . wait_for_log ( ' peer_out WIRE_REVOKE_AND_ACK ' )
l2 . daemon . wait_for_log ( ' peer_out WIRE_UPDATE_FULFILL_HTLC ' )
l1 . daemon . wait_for_log ( ' peer_in WIRE_REVOKE_AND_ACK ' )
2017-09-30 22:41:53 +02:00
2017-09-26 23:02:37 +02:00
# Payment should now complete.
t . result ( timeout = 10 )
# Now we really mess things up!
bitcoind . rpc . sendrawtransaction ( tx )
2017-11-28 05:39:22 +01:00
bitcoind . generate_block ( 1 )
2017-09-26 23:02:37 +02:00
l2 . daemon . wait_for_log ( ' -> ONCHAIND_CHEATED ' )
# FIXME: l1 should try to stumble along!
# l2 should spend all of the outputs (except to-us).
# Could happen in any order, depending on commitment tx.
l2 . daemon . wait_for_logs ( [ ' Propose handling THEIR_REVOKED_UNILATERAL/DELAYED_OUTPUT_TO_THEM by OUR_PENALTY_TX .* in 0 blocks ' ,
' sendrawtx exit 0 ' ,
' Propose handling THEIR_REVOKED_UNILATERAL/THEIR_HTLC by OUR_PENALTY_TX .* in 0 blocks ' ,
' sendrawtx exit 0 ' ] )
# FIXME: test HTLC tx race!
# 100 blocks later, all resolved.
2017-11-28 05:39:22 +01:00
bitcoind . generate_block ( 100 )
2017-09-26 23:02:37 +02:00
# FIXME: Test wallet balance...
l2 . daemon . wait_for_log ( ' onchaind complete, forgetting peer ' )
2017-09-30 19:37:14 +02:00
@unittest.skip ( " flaky test causing CI fails too often " )
2017-09-26 23:02:37 +02:00
def test_penalty_outhtlc ( self ) :
""" Test penalty transaction with an outgoing HTLC """
# First we need to get funds to l2, so suppress after second.
2017-10-11 12:08:50 +02:00
l1 = self . node_factory . get_node ( disconnect = [ ' =WIRE_COMMITMENT_SIGNED*3-nocommit ' ] , may_fail = True )
l2 = self . node_factory . get_node ( disconnect = [ ' =WIRE_COMMITMENT_SIGNED*3-nocommit ' ] )
2017-09-26 23:02:37 +02:00
2017-12-04 10:12:22 +01:00
l1 . rpc . connect ( l2 . info [ ' id ' ] , ' localhost ' , l2 . info [ ' port ' ] )
2017-09-26 23:02:37 +02:00
self . fund_channel ( l1 , l2 , 10 * * 6 )
# Move some across to l2.
self . pay ( l1 , l2 , 200000000 )
2017-10-11 12:08:50 +02:00
assert not l1 . daemon . is_in_log ( ' =WIRE_COMMITMENT_SIGNED ' )
assert not l2 . daemon . is_in_log ( ' =WIRE_COMMITMENT_SIGNED ' )
2017-09-26 23:02:37 +02:00
# Now, this will get stuck due to l1 commit being disabled..
t = self . pay ( l2 , l1 , 100000000 , async = True )
# Make sure we get signature from them.
l1 . daemon . wait_for_log ( ' peer_in WIRE_UPDATE_ADD_HTLC ' )
l1 . daemon . wait_for_log ( ' peer_in WIRE_COMMITMENT_SIGNED ' )
# They should both have commitments blocked now.
2017-10-11 12:08:50 +02:00
l1 . daemon . wait_for_log ( ' dev_disconnect: =WIRE_COMMITMENT_SIGNED ' )
l2 . daemon . wait_for_log ( ' dev_disconnect: =WIRE_COMMITMENT_SIGNED ' )
2017-09-26 23:02:37 +02:00
# Take our snapshot.
tx = l1 . rpc . dev_sign_last_tx ( l2 . info [ ' id ' ] ) [ ' tx ' ]
# Let the continue
l1 . rpc . dev_reenable_commit ( l2 . info [ ' id ' ] )
l2 . rpc . dev_reenable_commit ( l1 . info [ ' id ' ] )
# Thread should complete.
t . result ( timeout = 10 )
# Make sure both sides got revoke_and_ack for final.
l1 . daemon . wait_for_log ( ' peer_in WIRE_REVOKE_AND_ACK ' )
l2 . daemon . wait_for_log ( ' peer_in WIRE_REVOKE_AND_ACK ' )
# Now we really mess things up!
bitcoind . rpc . sendrawtransaction ( tx )
2017-11-28 05:39:22 +01:00
bitcoind . generate_block ( 1 )
2017-09-26 23:02:37 +02:00
l2 . daemon . wait_for_log ( ' -> ONCHAIND_CHEATED ' )
# FIXME: l1 should try to stumble along!
# l2 should spend all of the outputs (except to-us).
# Could happen in any order, depending on commitment tx.
l2 . daemon . wait_for_logs ( [ ' Ignoring output.*: THEIR_REVOKED_UNILATERAL/OUTPUT_TO_US ' ,
' Propose handling THEIR_REVOKED_UNILATERAL/DELAYED_OUTPUT_TO_THEM by OUR_PENALTY_TX .* in 0 blocks ' ,
' sendrawtx exit 0 ' ,
' Propose handling THEIR_REVOKED_UNILATERAL/OUR_HTLC by OUR_PENALTY_TX .* in 0 blocks ' ,
' sendrawtx exit 0 ' ] )
# FIXME: test HTLC tx race!
# 100 blocks later, all resolved.
2017-11-28 05:39:22 +01:00
bitcoind . generate_block ( 100 )
2017-09-26 23:02:37 +02:00
# FIXME: Test wallet balance...
l2 . daemon . wait_for_log ( ' onchaind complete, forgetting peer ' )
2017-10-24 04:06:14 +02:00
@unittest.skipIf ( not DEVELOPER , " needs DEVELOPER=1 " )
2017-08-23 03:55:27 +02:00
def test_permfail_new_commit ( self ) :
# Test case where we have two possible commits: it will use new one.
disconnects = [ ' -WIRE_REVOKE_AND_ACK ' , ' permfail ' ]
2017-08-28 17:53:01 +02:00
l1 = self . node_factory . get_node ( )
l2 = self . node_factory . get_node ( disconnect = disconnects )
2017-08-23 03:55:27 +02:00
2017-12-04 10:12:22 +01:00
l1 . rpc . connect ( l2 . info [ ' id ' ] , ' localhost ' , l2 . info [ ' port ' ] )
2017-08-23 03:55:27 +02:00
self . fund_channel ( l1 , l2 , 10 * * 6 )
# This will fail at l2's end.
t = self . pay ( l1 , l2 , 200000000 , async = True )
2017-09-30 22:41:53 +02:00
2017-08-23 03:55:27 +02:00
l2 . daemon . wait_for_log ( ' dev_disconnect permfail ' )
l2 . daemon . wait_for_log ( ' sendrawtx exit 0 ' )
2017-11-28 05:39:22 +01:00
bitcoind . generate_block ( 1 )
2017-08-23 03:55:27 +02:00
l1 . daemon . wait_for_log ( ' Their unilateral tx, new commit point ' )
l1 . daemon . wait_for_log ( ' -> ONCHAIND_THEIR_UNILATERAL ' )
l2 . daemon . wait_for_log ( ' -> ONCHAIND_OUR_UNILATERAL ' )
l2 . daemon . wait_for_log ( ' Propose handling OUR_UNILATERAL/THEIR_HTLC by THEIR_HTLC_TIMEOUT_TO_THEM \\ (IGNORING \\ ) in 5 blocks ' )
l1 . daemon . wait_for_log ( ' Propose handling THEIR_UNILATERAL/OUR_HTLC by OUR_HTLC_TIMEOUT_TO_US (.*) in 5 blocks ' )
# OK, time out HTLC.
2017-11-28 05:39:22 +01:00
bitcoind . generate_block ( 5 )
2017-08-23 03:55:27 +02:00
l1 . daemon . wait_for_log ( ' sendrawtx exit 0 ' )
2017-11-28 05:39:22 +01:00
bitcoind . generate_block ( 1 )
2017-08-23 03:55:27 +02:00
l1 . daemon . wait_for_log ( ' Resolved THEIR_UNILATERAL/OUR_HTLC by our proposal OUR_HTLC_TIMEOUT_TO_US ' )
l2 . daemon . wait_for_log ( ' Ignoring output.*: OUR_UNILATERAL/THEIR_HTLC ' )
t . cancel ( )
# Now, 100 blocks it should be done.
2017-11-28 05:39:22 +01:00
bitcoind . generate_block ( 100 )
2017-08-23 03:55:27 +02:00
l1 . daemon . wait_for_log ( ' onchaind complete, forgetting peer ' )
l2 . daemon . wait_for_log ( ' onchaind complete, forgetting peer ' )
2017-09-30 22:41:53 +02:00
2017-10-24 04:06:14 +02:00
@unittest.skipIf ( not DEVELOPER , " needs DEVELOPER=1 " )
2017-08-23 03:55:23 +02:00
def test_permfail_htlc_in ( self ) :
# Test case where we fail with unsettled incoming HTLC.
disconnects = [ ' -WIRE_UPDATE_FULFILL_HTLC ' , ' permfail ' ]
2017-08-28 17:53:01 +02:00
l1 = self . node_factory . get_node ( )
l2 = self . node_factory . get_node ( disconnect = disconnects )
2017-08-23 03:55:23 +02:00
2017-12-04 10:12:22 +01:00
l1 . rpc . connect ( l2 . info [ ' id ' ] , ' localhost ' , l2 . info [ ' port ' ] )
2017-08-23 03:55:23 +02:00
self . fund_channel ( l1 , l2 , 10 * * 6 )
# This will fail at l2's end.
2017-09-07 21:33:55 +02:00
t = self . pay ( l1 , l2 , 200000000 , async = True )
2017-08-23 03:55:23 +02:00
l2 . daemon . wait_for_log ( ' dev_disconnect permfail ' )
l2 . daemon . wait_for_log ( ' sendrawtx exit 0 ' )
2017-11-28 05:39:22 +01:00
bitcoind . generate_block ( 1 )
2017-08-23 03:55:23 +02:00
l1 . daemon . wait_for_log ( ' Their unilateral tx, old commit point ' )
l1 . daemon . wait_for_log ( ' -> ONCHAIND_THEIR_UNILATERAL ' )
l2 . daemon . wait_for_log ( ' -> ONCHAIND_OUR_UNILATERAL ' )
l2 . daemon . wait_for_log ( ' Propose handling OUR_UNILATERAL/THEIR_HTLC by THEIR_HTLC_TIMEOUT_TO_THEM \\ (IGNORING \\ ) in 5 blocks ' )
l1 . daemon . wait_for_log ( ' Propose handling THEIR_UNILATERAL/OUR_HTLC by OUR_HTLC_TIMEOUT_TO_US (.*) in 5 blocks ' )
2017-09-26 06:57:19 +02:00
# l2 then gets preimage, uses it instead of ignoring
2017-09-20 06:45:41 +02:00
l2 . daemon . wait_for_log ( ' Propose handling OUR_UNILATERAL/THEIR_HTLC by OUR_HTLC_SUCCESS_TX .* in 0 blocks ' )
2017-09-26 06:57:19 +02:00
l2 . daemon . wait_for_log ( ' sendrawtx exit 0 ' )
2017-11-28 05:39:22 +01:00
bitcoind . generate_block ( 1 )
2017-09-26 06:57:19 +02:00
# OK, l1 sees l2 fulfill htlc.
l1 . daemon . wait_for_log ( ' THEIR_UNILATERAL/OUR_HTLC gave us preimage ' )
2017-10-23 06:16:57 +02:00
l2 . daemon . wait_for_log ( ' Propose handling OUR_HTLC_SUCCESS_TX/DELAYED_OUTPUT_TO_US by OUR_DELAYED_RETURN_TO_WALLET .* in 5 blocks ' )
2017-11-28 05:39:22 +01:00
bitcoind . generate_block ( 5 )
2017-08-23 03:55:23 +02:00
2017-09-26 06:57:19 +02:00
l2 . daemon . wait_for_log ( ' sendrawtx exit 0 ' )
2017-09-30 22:41:53 +02:00
2017-08-23 03:55:23 +02:00
t . cancel ( )
# Now, 100 blocks it should be done.
2017-11-28 05:39:22 +01:00
bitcoind . generate_block ( 95 )
2017-09-26 06:57:19 +02:00
l1 . daemon . wait_for_log ( ' onchaind complete, forgetting peer ' )
assert not l2 . daemon . is_in_log ( ' onchaind complete, forgetting peer ' )
2017-11-28 05:39:22 +01:00
bitcoind . generate_block ( 5 )
2017-08-23 03:55:23 +02:00
l2 . daemon . wait_for_log ( ' onchaind complete, forgetting peer ' )
2017-09-30 22:41:53 +02:00
2017-10-24 04:06:14 +02:00
@unittest.skipIf ( not DEVELOPER , " needs DEVELOPER=1 " )
2017-08-23 03:55:23 +02:00
def test_permfail_htlc_out ( self ) :
# Test case where we fail with unsettled outgoing HTLC.
disconnects = [ ' +WIRE_REVOKE_AND_ACK ' , ' permfail ' ]
2017-08-28 17:53:01 +02:00
l1 = self . node_factory . get_node ( )
l2 = self . node_factory . get_node ( disconnect = disconnects )
2017-08-23 03:55:23 +02:00
2017-12-04 10:12:22 +01:00
l1 . rpc . connect ( l2 . info [ ' id ' ] , ' localhost ' , l2 . info [ ' port ' ] )
2017-08-23 03:55:23 +02:00
self . fund_channel ( l2 , l1 , 10 * * 6 )
# This will fail at l2's end.
t = self . pay ( l2 , l1 , 200000000 , async = True )
l2 . daemon . wait_for_log ( ' dev_disconnect permfail ' )
l2 . daemon . wait_for_log ( ' sendrawtx exit 0 ' )
2017-11-28 05:39:22 +01:00
bitcoind . generate_block ( 1 )
2017-08-23 03:55:23 +02:00
l1 . daemon . wait_for_log ( ' Their unilateral tx, old commit point ' )
l1 . daemon . wait_for_log ( ' -> ONCHAIND_THEIR_UNILATERAL ' )
l2 . daemon . wait_for_log ( ' -> ONCHAIND_OUR_UNILATERAL ' )
2017-09-26 06:57:19 +02:00
l2 . daemon . wait_for_logs ( [ ' Propose handling OUR_UNILATERAL/OUR_HTLC by OUR_HTLC_TIMEOUT_TO_US \\ (.* \\ ) in 5 blocks ' ,
2017-10-23 06:16:57 +02:00
' Propose handling OUR_UNILATERAL/DELAYED_OUTPUT_TO_US by OUR_DELAYED_RETURN_TO_WALLET .* in 5 blocks ' ] )
2017-08-23 03:55:23 +02:00
2017-09-26 06:57:19 +02:00
l1 . daemon . wait_for_log ( ' Propose handling THEIR_UNILATERAL/THEIR_HTLC by THEIR_HTLC_TIMEOUT_TO_THEM \\ (IGNORING \\ ) in 5 blocks ' )
# l1 then gets preimage, uses it instead of ignoring
l1 . daemon . wait_for_log ( ' Propose handling THEIR_UNILATERAL/THEIR_HTLC by THEIR_HTLC_FULFILL_TO_US .* in 0 blocks ' )
l1 . daemon . wait_for_log ( ' sendrawtx exit 0 ' )
2017-09-20 06:45:41 +02:00
2017-09-26 06:57:19 +02:00
# l2 sees l1 fulfill tx.
2017-11-28 05:39:22 +01:00
bitcoind . generate_block ( 1 )
2017-08-23 03:55:23 +02:00
2017-09-26 06:57:19 +02:00
l2 . daemon . wait_for_log ( ' OUR_UNILATERAL/OUR_HTLC gave us preimage ' )
2017-08-23 03:55:23 +02:00
t . cancel ( )
2017-10-23 06:16:57 +02:00
# l2 can send OUR_DELAYED_RETURN_TO_WALLET after 4 more blocks.
2017-11-28 05:39:22 +01:00
bitcoind . generate_block ( 4 )
2017-09-26 06:57:19 +02:00
l2 . daemon . wait_for_log ( ' Broadcasting OUR_DELAYED_RETURN_TO_WALLET .* to resolve OUR_UNILATERAL/DELAYED_OUTPUT_TO_US ' )
l2 . daemon . wait_for_log ( ' sendrawtx exit 0 ' )
# Now, 100 blocks they should be done.
2017-11-28 05:39:22 +01:00
bitcoind . generate_block ( 94 )
2017-09-26 06:57:19 +02:00
assert not l1 . daemon . is_in_log ( ' onchaind complete, forgetting peer ' )
assert not l2 . daemon . is_in_log ( ' onchaind complete, forgetting peer ' )
2017-11-28 05:39:22 +01:00
bitcoind . generate_block ( 1 )
2017-08-23 03:55:23 +02:00
l1 . daemon . wait_for_log ( ' onchaind complete, forgetting peer ' )
2017-09-26 06:57:19 +02:00
assert not l2 . daemon . is_in_log ( ' onchaind complete, forgetting peer ' )
2017-11-28 05:39:22 +01:00
bitcoind . generate_block ( 5 )
2017-09-26 06:57:19 +02:00
assert not l2 . daemon . is_in_log ( ' onchaind complete, forgetting peer ' )
2017-11-28 05:39:22 +01:00
bitcoind . generate_block ( 1 )
2017-09-26 06:57:19 +02:00
l2 . daemon . wait_for_log ( ' onchaind complete, forgetting peer ' )
2017-08-23 03:55:23 +02:00
2017-04-29 03:02:26 +02:00
def test_gossip_jsonrpc ( self ) :
l1 , l2 = self . connect ( )
self . fund_channel ( l1 , l2 , 10 * * 5 )
2017-06-06 01:47:10 +02:00
# Shouldn't send announce signatures until 6 deep.
assert not l1 . daemon . is_in_log ( ' peer_out WIRE_ANNOUNCEMENT_SIGNATURES ' )
2017-11-28 05:39:22 +01:00
l1 . bitcoin . generate_block ( 5 )
2017-07-01 09:11:59 +02:00
# Could happen in either order.
l1 . daemon . wait_for_logs ( [ ' peer_out WIRE_ANNOUNCEMENT_SIGNATURES ' ,
' peer_in WIRE_ANNOUNCEMENT_SIGNATURES ' ] )
2017-04-04 13:44:25 +02:00
2017-07-01 09:11:59 +02:00
# Could happen in either order.
l1 . daemon . wait_for_logs ( [ ' peer_out WIRE_CHANNEL_ANNOUNCEMENT ' ,
' peer_in WIRE_CHANNEL_ANNOUNCEMENT ' ] )
2017-03-15 21:52:49 +01:00
nodes = l1 . rpc . getnodes ( ) [ ' nodes ' ]
assert set ( [ n [ ' nodeid ' ] for n in nodes ] ) == set ( [ l1 . info [ ' id ' ] , l2 . info [ ' id ' ] ] )
2017-05-02 14:32:44 +02:00
l1 . daemon . wait_for_log ( ' peer_in WIRE_CHANNEL_UPDATE ' )
l2 . daemon . wait_for_log ( ' peer_in WIRE_CHANNEL_UPDATE ' )
2017-04-04 13:44:25 +02:00
channels = l1 . rpc . getchannels ( ) [ ' channels ' ]
assert len ( channels ) == 2
2017-11-23 13:08:36 +01:00
wait_for ( lambda : [ c [ ' active ' ] for c in channels ] == [ True , True ] )
2017-04-04 13:44:25 +02:00
2017-04-29 03:01:44 +02:00
def ping_tests ( self , l1 , l2 ) :
# 0-byte pong gives just type + length field.
ret = l1 . rpc . dev_ping ( l2 . info [ ' id ' ] , 0 , 0 )
assert ret [ ' totlen ' ] == 4
# 1000-byte ping, 0-byte pong.
ret = l1 . rpc . dev_ping ( l2 . info [ ' id ' ] , 1000 , 0 )
assert ret [ ' totlen ' ] == 4
# 1000 byte pong.
ret = l1 . rpc . dev_ping ( l2 . info [ ' id ' ] , 1000 , 1000 )
assert ret [ ' totlen ' ] == 1004
# Maximum length pong.
ret = l1 . rpc . dev_ping ( l2 . info [ ' id ' ] , 1000 , 65531 )
assert ret [ ' totlen ' ] == 65535
# Overlength -> no reply.
for s in range ( 65532 , 65536 ) :
ret = l1 . rpc . dev_ping ( l2 . info [ ' id ' ] , 1000 , s )
assert ret [ ' totlen ' ] == 0
2017-10-24 04:06:14 +02:00
@unittest.skipIf ( not DEVELOPER , " needs DEVELOPER=1 " )
2017-04-29 03:01:44 +02:00
def test_ping ( self ) :
2017-04-29 03:02:26 +02:00
l1 , l2 = self . connect ( )
2017-04-29 03:01:44 +02:00
# Test gossip pinging.
self . ping_tests ( l1 , l2 )
2017-04-29 03:02:26 +02:00
self . fund_channel ( l1 , l2 , 10 * * 5 )
2017-04-29 03:01:44 +02:00
# channeld pinging
self . ping_tests ( l1 , l2 )
2017-10-24 04:06:14 +02:00
@unittest.skipIf ( not DEVELOPER , " needs DEVELOPER=1 " )
2017-07-04 20:36:24 +02:00
def test_routing_gossip_reconnect ( self ) :
# Connect two peers, reconnect and then see if we resume the
# gossip.
disconnects = [ ' -WIRE_CHANNEL_ANNOUNCEMENT ' ]
2017-08-28 17:53:01 +02:00
l1 = self . node_factory . get_node ( disconnect = disconnects )
l2 = self . node_factory . get_node ( )
l3 = self . node_factory . get_node ( )
2017-12-04 10:12:22 +01:00
l1 . rpc . connect ( l2 . info [ ' id ' ] , ' localhost ' , l2 . info [ ' port ' ] )
2017-07-04 20:36:24 +02:00
l1 . openchannel ( l2 , 20000 )
# Now open new channels and everybody should sync
2017-12-04 10:12:22 +01:00
l2 . rpc . connect ( l3 . info [ ' id ' ] , ' localhost ' , l3 . info [ ' port ' ] )
2017-07-04 20:36:24 +02:00
l2 . openchannel ( l3 , 20000 )
# Settle the gossip
for n in [ l1 , l2 , l3 ] :
wait_for ( lambda : len ( n . rpc . getchannels ( ) [ ' channels ' ] ) == 4 )
2017-09-12 13:50:47 +02:00
def test_second_channel ( self ) :
l1 = self . node_factory . get_node ( )
l2 = self . node_factory . get_node ( )
l3 = self . node_factory . get_node ( )
2017-12-04 10:12:22 +01:00
l1 . rpc . connect ( l2 . info [ ' id ' ] , ' localhost ' , l2 . info [ ' port ' ] )
l1 . rpc . connect ( l3 . info [ ' id ' ] , ' localhost ' , l3 . info [ ' port ' ] )
2017-09-12 13:50:47 +02:00
self . fund_channel ( l1 , l2 , 10 * * 6 )
self . fund_channel ( l1 , l3 , 10 * * 6 )
2017-10-24 04:06:14 +02:00
@unittest.skipIf ( not DEVELOPER , " needs DEVELOPER=1 for --dev-broadcast-interval " )
2017-04-24 20:27:53 +02:00
def test_routing_gossip ( self ) :
2017-08-28 17:53:01 +02:00
nodes = [ self . node_factory . get_node ( ) for _ in range ( 5 ) ]
2017-04-24 20:27:53 +02:00
l1 = nodes [ 0 ]
l5 = nodes [ 4 ]
for i in range ( len ( nodes ) - 1 ) :
src , dst = nodes [ i ] , nodes [ i + 1 ]
2017-12-04 10:12:22 +01:00
src . rpc . connect ( dst . info [ ' id ' ] , ' localhost ' , dst . info [ ' port ' ] )
2017-04-24 20:27:53 +02:00
src . openchannel ( dst , 20000 )
2017-06-06 01:47:10 +02:00
# Allow announce messages.
2017-11-28 05:39:22 +01:00
l1 . bitcoin . generate_block ( 5 )
2017-06-06 01:47:10 +02:00
2017-04-24 20:27:53 +02:00
def settle_gossip ( n ) :
""" Wait for gossip to settle at the node
"""
expected_connections = 2 * ( len ( nodes ) - 1 )
start_time = time . time ( )
# Wait at most 10 seconds, broadcast interval is 1 second
while time . time ( ) - start_time < 10 :
channels = n . rpc . getchannels ( ) [ ' channels ' ]
if len ( channels ) == expected_connections :
break
else :
time . sleep ( 0.1 )
for n in nodes :
settle_gossip ( n )
# Deep check that all channels are in there
comb = [ ]
for i in range ( len ( nodes ) - 1 ) :
comb . append ( ( nodes [ i ] . info [ ' id ' ] , nodes [ i + 1 ] . info [ ' id ' ] ) )
comb . append ( ( nodes [ i + 1 ] . info [ ' id ' ] , nodes [ i ] . info [ ' id ' ] ) )
for n in nodes :
seen = [ ]
channels = n . rpc . getchannels ( ) [ ' channels ' ]
for c in channels :
seen . append ( ( c [ ' source ' ] , c [ ' destination ' ] ) )
assert set ( seen ) == set ( comb )
2017-04-27 08:08:50 +02:00
def test_forward ( self ) :
# Connect 1 -> 2 -> 3.
l1 , l2 = self . connect ( )
2017-08-28 17:53:01 +02:00
l3 = self . node_factory . get_node ( )
2017-12-04 10:12:22 +01:00
ret = l2 . rpc . connect ( l3 . info [ ' id ' ] , ' localhost ' , l3 . info [ ' port ' ] )
2017-04-27 08:08:50 +02:00
assert ret [ ' id ' ] == l3 . info [ ' id ' ]
gossipd: rewrite to do the handshake internally.
Now the flow is much simpler from a lightningd POV:
1. If we want to connect to a peer, just send gossipd `gossipctl_reach_peer`.
2. Every new peer, gossipd hands up to lightningd, with global/local features
and the peer fd and a gossip fd using `gossip_peer_connected`
3. If lightningd doesn't want it, it just hands the peerfd and global/local
features back to gossipd using `gossipctl_handle_peer`
4. If a peer sends a non-gossip msg (eg `open_channel`) the gossipd sends
it up using `gossip_peer_nongossip`.
5. If lightningd wants to fund a channel, it simply calls `release_channel`.
Notes:
* There's no more "unique_id": we use the peer id.
* For the moment, we don't ask gossipd when we're told to list peers, so
connected peers without a channel don't appear in the JSON getpeers API.
* We add a `gossipctl_peer_addrhint` for the moment, so you can connect to
a specific ip/port, but using other sources is a TODO.
* We now (correctly) only give up on reaching a peer after we exchange init
messages, which changes the test_disconnect case.
Signed-off-by: Rusty Russell <rusty@rustcorp.com.au>
2017-10-11 12:09:49 +02:00
l3 . daemon . wait_for_log ( ' WIRE_GOSSIPCTL_HANDLE_PEER ' )
2017-04-27 08:08:50 +02:00
self . fund_channel ( l1 , l2 , 10 * * 6 )
self . fund_channel ( l2 , l3 , 10 * * 6 )
2017-06-06 01:47:10 +02:00
# Allow announce messages.
2017-11-28 05:39:22 +01:00
l1 . bitcoin . generate_block ( 5 )
2017-08-14 22:36:42 +02:00
2017-04-27 08:08:50 +02:00
# If they're at different block heights we can get spurious errors.
2017-07-04 20:47:50 +02:00
sync_blockheight ( [ l1 , l2 , l3 ] )
2017-04-27 08:08:50 +02:00
2017-05-02 07:26:31 +02:00
chanid1 = l1 . rpc . getpeer ( l2 . info [ ' id ' ] ) [ ' channel ' ]
chanid2 = l2 . rpc . getpeer ( l3 . info [ ' id ' ] ) [ ' channel ' ]
assert l2 . rpc . getpeer ( l1 . info [ ' id ' ] ) [ ' channel ' ] == chanid1
assert l3 . rpc . getpeer ( l2 . info [ ' id ' ] ) [ ' channel ' ] == chanid2
2017-10-26 05:06:19 +02:00
rhash = l3 . rpc . invoice ( 100000000 , ' testpayment1 ' , ' desc ' ) [ ' rhash ' ]
2017-04-27 08:08:50 +02:00
assert l3 . rpc . listinvoice ( ' testpayment1 ' ) [ 0 ] [ ' complete ' ] == False
# Fee for node2 is 10 millionths, plus 1.
amt = 100000000
fee = amt * 10 / / 1000000 + 1
2017-05-02 07:26:31 +02:00
baseroute = [ { ' msatoshi ' : amt + fee ,
' id ' : l2 . info [ ' id ' ] ,
2017-09-06 22:47:29 +02:00
' delay ' : 12 ,
2017-05-02 07:26:31 +02:00
' channel ' : chanid1 } ,
{ ' msatoshi ' : amt ,
' id ' : l3 . info [ ' id ' ] ,
2017-09-06 22:47:29 +02:00
' delay ' : 6 ,
2017-05-02 07:26:31 +02:00
' channel ' : chanid2 } ]
2017-04-27 08:08:50 +02:00
# Unknown other peer
2017-05-02 07:26:31 +02:00
route = copy . deepcopy ( baseroute )
route [ 1 ] [ ' id ' ] = ' 031a8dc444e41bb989653a4501e11175a488a57439b0c4947704fd6e3de5dca607 '
2017-04-27 08:08:50 +02:00
self . assertRaises ( ValueError , l1 . rpc . sendpay , to_json ( route ) , rhash )
# Delay too short (we always add one internally anyway, so subtract 2 here).
2017-05-02 07:26:31 +02:00
route = copy . deepcopy ( baseroute )
route [ 0 ] [ ' delay ' ] = 8
2017-04-27 08:08:50 +02:00
self . assertRaises ( ValueError , l1 . rpc . sendpay , to_json ( route ) , rhash )
# Final delay too short
2017-05-02 07:26:31 +02:00
route = copy . deepcopy ( baseroute )
route [ 1 ] [ ' delay ' ] = 3
2017-04-27 08:08:50 +02:00
self . assertRaises ( ValueError , l1 . rpc . sendpay , to_json ( route ) , rhash )
2017-08-14 22:36:42 +02:00
2017-04-27 08:08:50 +02:00
# This one works
2017-05-02 07:26:31 +02:00
route = copy . deepcopy ( baseroute )
2017-04-27 08:08:50 +02:00
l1 . rpc . sendpay ( to_json ( route ) , rhash )
2017-04-24 20:27:53 +02:00
2017-10-23 06:16:57 +02:00
@unittest.skipIf ( not DEVELOPER , " needs DEVELOPER=1 for --dev-broadcast-interval " )
def test_forward_different_fees_and_cltv ( self ) :
# FIXME: Check BOLT quotes here too
# BOLT #7:
# ```
# B
# / \
# / \
2017-11-08 14:34:11 +01:00
# A C
2017-10-23 06:16:57 +02:00
# \ /
# \ /
# D
# ```
2017-11-08 14:34:11 +01:00
#
2017-10-23 06:16:57 +02:00
# Each advertises the following `cltv_expiry_delta` on its end of every
# channel:
#
# 1. A: 10 blocks
# 2. B: 20 blocks
# 3. C: 30 blocks
# 4. D: 40 blocks
#
# C also uses a minimum `cltv_expiry` of 9 (the default) when requesting
# payments.
#
# Also, each node has the same fee scheme which it uses for each of its
# channels:
#
# 1. A: 100 base + 1000 millionths
# 1. B: 200 base + 2000 millionths
# 1. C: 300 base + 3000 millionths
# 1. D: 400 base + 4000 millionths
# We don't do D yet.
l1 = self . node_factory . get_node ( options = [ ' --cltv-delta=10 ' , ' --fee-base=100 ' , ' --fee-per-satoshi=1000 ' ] )
l2 = self . node_factory . get_node ( options = [ ' --cltv-delta=20 ' , ' --fee-base=200 ' , ' --fee-per-satoshi=2000 ' ] )
l3 = self . node_factory . get_node ( options = [ ' --cltv-delta=30 ' , ' --cltv-final=9 ' , ' --fee-base=300 ' , ' --fee-per-satoshi=3000 ' ] )
2017-12-04 10:12:22 +01:00
ret = l1 . rpc . connect ( l2 . info [ ' id ' ] , ' localhost ' , l2 . info [ ' port ' ] )
2017-10-23 06:16:57 +02:00
assert ret [ ' id ' ] == l2 . info [ ' id ' ]
l1 . daemon . wait_for_log ( ' WIRE_GOSSIPCTL_HANDLE_PEER ' )
l2 . daemon . wait_for_log ( ' WIRE_GOSSIPCTL_HANDLE_PEER ' )
2017-12-04 10:12:22 +01:00
ret = l2 . rpc . connect ( l3 . info [ ' id ' ] , ' localhost ' , l3 . info [ ' port ' ] )
2017-10-23 06:16:57 +02:00
assert ret [ ' id ' ] == l3 . info [ ' id ' ]
l2 . daemon . wait_for_log ( ' WIRE_GOSSIPCTL_HANDLE_PEER ' )
l3 . daemon . wait_for_log ( ' WIRE_GOSSIPCTL_HANDLE_PEER ' )
c1 = self . fund_channel ( l1 , l2 , 10 * * 6 )
c2 = self . fund_channel ( l2 , l3 , 10 * * 6 )
# Allow announce messages.
2017-11-28 05:39:22 +01:00
l1 . bitcoin . generate_block ( 5 )
2017-10-23 06:16:57 +02:00
# Make sure l1 has seen announce for all channels.
l1 . daemon . wait_for_logs ( [
' Received channel_update for channel {} \\ (0 \\ ) ' . format ( c1 ) ,
' Received channel_update for channel {} \\ (1 \\ ) ' . format ( c1 ) ,
' Received channel_update for channel {} \\ (0 \\ ) ' . format ( c2 ) ,
' Received channel_update for channel {} \\ (1 \\ ) ' . format ( c2 ) ] )
# BOLT #7:
2017-11-08 14:34:11 +01:00
#
2017-10-23 06:16:57 +02:00
# If B were to send 4,999,999 millisatoshi directly to C, it wouldn't
# charge itself a fee nor add its own `cltv_expiry_delta`, so it would
# use C's requested `cltv_expiry` of 9. We also assume it adds a
# "shadow route" to give an extra CLTV of 42. It could also add extra
# cltv deltas at other hops, as these values are a minimum, but we don't
# here for simplicity:
# FIXME: Add shadow route
shadow_route = 0
route = l2 . rpc . getroute ( l3 . info [ ' id ' ] , 4999999 , 1 ) [ " route " ]
assert len ( route ) == 1
# BOLT #7:
#
# * `amount_msat`: 4999999
# * `cltv_expiry`: current-block-height + 9 + 42
# * `onion_routing_packet`:
# * `amt_to_forward` = 4999999
# * `outgoing_cltv_value` = current-block-height + 9 + 42
#
assert route [ 0 ] [ ' msatoshi ' ] == 4999999
assert route [ 0 ] [ ' delay ' ] == 9 + shadow_route
# BOLT #7:
# If A were to send an 4,999,999 millisatoshi to C via B, it needs to
# pay B the fee it specified in the B->C `channel_update`, calculated as
# per [HTLC Fees](#htlc_fees):
2017-11-08 14:34:11 +01:00
#
2017-10-23 06:16:57 +02:00
# 200 + 4999999 * 2000 / 1000000 = 10199
2017-11-08 14:34:11 +01:00
#
2017-10-23 06:16:57 +02:00
# Similarly, it would need to add the `cltv_expiry` from B->C's
# `channel_update` (20), plus C's requested minimum (9), plus 42 for the
# "shadow route". Thus the `update_add_htlc` message from A to B would
# be:
2017-11-08 14:34:11 +01:00
#
2017-10-23 06:16:57 +02:00
# * `amount_msat`: 5010198
# * `cltv_expiry`: current-block-height + 20 + 9 + 42
# * `onion_routing_packet`:
# * `amt_to_forward` = 4999999
# * `outgoing_cltv_value` = current-block-height + 9 + 42
route = l1 . rpc . getroute ( l3 . info [ ' id ' ] , 4999999 , 1 ) [ " route " ]
assert len ( route ) == 2
assert route [ 0 ] [ ' msatoshi ' ] == 5010198
assert route [ 0 ] [ ' delay ' ] == 20 + 9 + shadow_route
assert route [ 1 ] [ ' msatoshi ' ] == 4999999
assert route [ 1 ] [ ' delay ' ] == 9 + shadow_route
2017-10-26 05:06:19 +02:00
rhash = l3 . rpc . invoice ( 4999999 , ' test_forward_different_fees_and_cltv ' , ' desc ' ) [ ' rhash ' ]
2017-10-23 06:16:57 +02:00
assert l3 . rpc . listinvoice ( ' test_forward_different_fees_and_cltv ' ) [ 0 ] [ ' complete ' ] == False
# This should work.
l1 . rpc . sendpay ( to_json ( route ) , rhash )
# We add one to the blockcount for a bit of fuzz (FIXME: Shadowroute would fix this!)
shadow_route = 1
l1 . daemon . wait_for_log ( " Adding HTLC 0 msat=5010198 cltv= {} gave 0 "
. format ( bitcoind . rpc . getblockcount ( ) + 20 + 9 + shadow_route ) )
l2 . daemon . wait_for_log ( " Adding HTLC 0 msat=4999999 cltv= {} gave 0 "
. format ( bitcoind . rpc . getblockcount ( ) + 9 + shadow_route ) )
l3 . daemon . wait_for_log ( " test_forward_different_fees_and_cltv: Actual amount 4999999msat, HTLC expiry {} "
. format ( bitcoind . rpc . getblockcount ( ) + 9 + shadow_route ) )
assert l3 . rpc . listinvoice ( ' test_forward_different_fees_and_cltv ' ) [ 0 ] [ ' complete ' ] == True
2017-10-23 06:16:57 +02:00
@unittest.skipIf ( not DEVELOPER , " needs DEVELOPER=1 for --dev-broadcast-interval " )
def test_forward_pad_fees_and_cltv ( self ) :
""" Test that we are allowed extra locktime delta, and fees """
l1 = self . node_factory . get_node ( options = [ ' --cltv-delta=10 ' , ' --fee-base=100 ' , ' --fee-per-satoshi=1000 ' ] )
l2 = self . node_factory . get_node ( options = [ ' --cltv-delta=20 ' , ' --fee-base=200 ' , ' --fee-per-satoshi=2000 ' ] )
l3 = self . node_factory . get_node ( options = [ ' --cltv-delta=30 ' , ' --cltv-final=9 ' , ' --fee-base=300 ' , ' --fee-per-satoshi=3000 ' ] )
2017-12-04 10:12:22 +01:00
ret = l1 . rpc . connect ( l2 . info [ ' id ' ] , ' localhost ' , l2 . info [ ' port ' ] )
2017-10-23 06:16:57 +02:00
assert ret [ ' id ' ] == l2 . info [ ' id ' ]
l1 . daemon . wait_for_log ( ' WIRE_GOSSIPCTL_HANDLE_PEER ' )
l2 . daemon . wait_for_log ( ' WIRE_GOSSIPCTL_HANDLE_PEER ' )
2017-12-04 10:12:22 +01:00
ret = l2 . rpc . connect ( l3 . info [ ' id ' ] , ' localhost ' , l3 . info [ ' port ' ] )
2017-10-23 06:16:57 +02:00
assert ret [ ' id ' ] == l3 . info [ ' id ' ]
l2 . daemon . wait_for_log ( ' WIRE_GOSSIPCTL_HANDLE_PEER ' )
l3 . daemon . wait_for_log ( ' WIRE_GOSSIPCTL_HANDLE_PEER ' )
2017-11-08 14:34:11 +01:00
2017-10-23 06:16:57 +02:00
c1 = self . fund_channel ( l1 , l2 , 10 * * 6 )
c2 = self . fund_channel ( l2 , l3 , 10 * * 6 )
# Allow announce messages.
2017-11-28 05:39:22 +01:00
l1 . bitcoin . generate_block ( 5 )
2017-10-23 06:16:57 +02:00
# Make sure l1 has seen announce for all channels.
l1 . daemon . wait_for_logs ( [
' Received channel_update for channel {} \\ (0 \\ ) ' . format ( c1 ) ,
' Received channel_update for channel {} \\ (1 \\ ) ' . format ( c1 ) ,
' Received channel_update for channel {} \\ (0 \\ ) ' . format ( c2 ) ,
' Received channel_update for channel {} \\ (1 \\ ) ' . format ( c2 ) ] )
route = l1 . rpc . getroute ( l3 . info [ ' id ' ] , 4999999 , 1 ) [ " route " ]
assert len ( route ) == 2
assert route [ 0 ] [ ' msatoshi ' ] == 5010198
assert route [ 0 ] [ ' delay ' ] == 20 + 9
assert route [ 1 ] [ ' msatoshi ' ] == 4999999
assert route [ 1 ] [ ' delay ' ] == 9
# Modify so we overpay, overdo the cltv.
route [ 0 ] [ ' msatoshi ' ] + = 2000
route [ 0 ] [ ' delay ' ] + = 20
route [ 1 ] [ ' msatoshi ' ] + = 1000
route [ 1 ] [ ' delay ' ] + = 10
# This should work.
2017-10-26 05:06:19 +02:00
rhash = l3 . rpc . invoice ( 4999999 , ' test_forward_pad_fees_and_cltv ' , ' desc ' ) [ ' rhash ' ]
2017-10-23 06:16:57 +02:00
l1 . rpc . sendpay ( to_json ( route ) , rhash )
assert l3 . rpc . listinvoice ( ' test_forward_pad_fees_and_cltv ' ) [ 0 ] [ ' complete ' ] == True
2017-11-02 00:54:00 +01:00
@unittest.skipIf ( not DEVELOPER , " needs DEVELOPER=1 " )
def test_htlc_out_timeout ( self ) :
""" Test that we drop onchain if the peer doesn ' t time out HTLC """
# HTLC 1->2, 1 fails after it's irrevocably committed, can't reconnect
disconnects = [ ' @WIRE_REVOKE_AND_ACK ' ]
l1 = self . node_factory . get_node ( disconnect = disconnects ,
options = [ ' --no-reconnect ' ] )
l2 = self . node_factory . get_node ( )
2017-12-04 10:12:22 +01:00
l1 . rpc . connect ( l2 . info [ ' id ' ] , ' localhost ' , l2 . info [ ' port ' ] )
2017-11-02 00:54:00 +01:00
chanid = self . fund_channel ( l1 , l2 , 10 * * 6 )
# Wait for route propagation.
2017-11-28 05:39:22 +01:00
bitcoind . generate_block ( 5 )
2017-11-02 00:54:00 +01:00
l1 . daemon . wait_for_logs ( [ ' Received channel_update for channel {} \ (0 \ ) '
. format ( chanid ) ,
' Received channel_update for channel {} \ (1 \ ) '
. format ( chanid ) ] )
amt = 200000000
inv = l2 . rpc . invoice ( amt , ' test_htlc_out_timeout ' , ' desc ' ) [ ' bolt11 ' ]
assert l2 . rpc . listinvoice ( ' test_htlc_out_timeout ' ) [ 0 ] [ ' complete ' ] == False
payfuture = self . executor . submit ( l1 . rpc . pay , inv ) ;
# l1 will drop to chain, not reconnect.
l1 . daemon . wait_for_log ( ' dev_disconnect: @WIRE_REVOKE_AND_ACK ' )
# Takes 6 blocks to timeout (cltv-final + 1), but we also give grace period of 1 block.
2017-11-28 05:39:22 +01:00
bitcoind . generate_block ( 5 + 1 )
2017-11-02 00:54:00 +01:00
assert not l1 . daemon . is_in_log ( ' hit deadline ' )
2017-11-28 05:39:22 +01:00
bitcoind . generate_block ( 2 )
2017-11-02 00:54:00 +01:00
2017-11-08 14:34:11 +01:00
l1 . daemon . wait_for_log ( ' Offered HTLC 0 SENT_ADD_ACK_REVOCATION cltv .* hit deadline ' )
2017-11-02 00:54:00 +01:00
l1 . daemon . wait_for_log ( ' sendrawtx exit 0 ' )
2017-11-28 05:39:22 +01:00
l1 . bitcoin . generate_block ( 1 )
2017-11-02 00:54:00 +01:00
l1 . daemon . wait_for_log ( ' -> ONCHAIND_OUR_UNILATERAL ' )
l2 . daemon . wait_for_log ( ' -> ONCHAIND_THEIR_UNILATERAL ' )
2017-11-02 00:56:07 +01:00
@unittest.skipIf ( not DEVELOPER , " needs DEVELOPER=1 " )
def test_htlc_in_timeout ( self ) :
""" Test that we drop onchain if the peer doesn ' t accept fulfilled HTLC """
# HTLC 1->2, 1 fails after 2 has sent committed the fulfill
disconnects = [ ' -WIRE_REVOKE_AND_ACK*2 ' ]
l1 = self . node_factory . get_node ( disconnect = disconnects ,
options = [ ' --no-reconnect ' ] )
l2 = self . node_factory . get_node ( )
2017-12-04 10:12:22 +01:00
l1 . rpc . connect ( l2 . info [ ' id ' ] , ' localhost ' , l2 . info [ ' port ' ] )
2017-11-02 00:56:07 +01:00
chanid = self . fund_channel ( l1 , l2 , 10 * * 6 )
# Wait for route propagation.
2017-11-28 05:39:22 +01:00
bitcoind . generate_block ( 5 )
2017-11-02 00:56:07 +01:00
l1 . daemon . wait_for_logs ( [ ' Received channel_update for channel {} \ (0 \ ) '
. format ( chanid ) ,
' Received channel_update for channel {} \ (1 \ ) '
. format ( chanid ) ] )
amt = 200000000
inv = l2 . rpc . invoice ( amt , ' test_htlc_in_timeout ' , ' desc ' ) [ ' bolt11 ' ]
assert l2 . rpc . listinvoice ( ' test_htlc_in_timeout ' ) [ 0 ] [ ' complete ' ] == False
payfuture = self . executor . submit ( l1 . rpc . pay , inv ) ;
# l1 will drop to chain, not reconnect.
l1 . daemon . wait_for_log ( ' dev_disconnect: -WIRE_REVOKE_AND_ACK ' )
2017-11-08 14:34:11 +01:00
# Deadline HTLC expiry minus 1/2 cltv-expiry delta (rounded up) (== cltv - 3). ctlv is 5+1.
2017-11-28 05:39:22 +01:00
bitcoind . generate_block ( 2 )
2017-11-02 00:56:07 +01:00
assert not l2 . daemon . is_in_log ( ' hit deadline ' )
2017-11-28 05:39:22 +01:00
bitcoind . generate_block ( 2 )
2017-11-02 00:56:07 +01:00
2017-11-08 14:34:11 +01:00
l2 . daemon . wait_for_log ( ' Fulfilled HTLC 0 SENT_REMOVE_COMMIT cltv .* hit deadline ' )
2017-11-02 00:56:07 +01:00
l2 . daemon . wait_for_log ( ' sendrawtx exit 0 ' )
2017-11-28 05:39:22 +01:00
l2 . bitcoin . generate_block ( 1 )
2017-11-02 00:56:07 +01:00
l2 . daemon . wait_for_log ( ' -> ONCHAIND_OUR_UNILATERAL ' )
l1 . daemon . wait_for_log ( ' -> ONCHAIND_THEIR_UNILATERAL ' )
2017-10-24 04:06:14 +02:00
@unittest.skipIf ( not DEVELOPER , " needs DEVELOPER=1 " )
2017-05-25 04:15:57 +02:00
def test_disconnect ( self ) :
gossipd: rewrite to do the handshake internally.
Now the flow is much simpler from a lightningd POV:
1. If we want to connect to a peer, just send gossipd `gossipctl_reach_peer`.
2. Every new peer, gossipd hands up to lightningd, with global/local features
and the peer fd and a gossip fd using `gossip_peer_connected`
3. If lightningd doesn't want it, it just hands the peerfd and global/local
features back to gossipd using `gossipctl_handle_peer`
4. If a peer sends a non-gossip msg (eg `open_channel`) the gossipd sends
it up using `gossip_peer_nongossip`.
5. If lightningd wants to fund a channel, it simply calls `release_channel`.
Notes:
* There's no more "unique_id": we use the peer id.
* For the moment, we don't ask gossipd when we're told to list peers, so
connected peers without a channel don't appear in the JSON getpeers API.
* We add a `gossipctl_peer_addrhint` for the moment, so you can connect to
a specific ip/port, but using other sources is a TODO.
* We now (correctly) only give up on reaching a peer after we exchange init
messages, which changes the test_disconnect case.
Signed-off-by: Rusty Russell <rusty@rustcorp.com.au>
2017-10-11 12:09:49 +02:00
# These should all make us fail, and retry.
# FIXME: Configure short timeout for reconnect!
2017-05-25 04:15:57 +02:00
disconnects = [ ' -WIRE_INIT ' ,
' @WIRE_INIT ' ,
' +WIRE_INIT ' ]
2017-08-28 17:53:01 +02:00
l1 = self . node_factory . get_node ( disconnect = disconnects )
l2 = self . node_factory . get_node ( )
2017-12-04 10:12:22 +01:00
l1 . rpc . connect ( l2 . info [ ' id ' ] , ' localhost ' , l2 . info [ ' port ' ] )
2017-05-25 04:15:57 +02:00
gossipd: rewrite to do the handshake internally.
Now the flow is much simpler from a lightningd POV:
1. If we want to connect to a peer, just send gossipd `gossipctl_reach_peer`.
2. Every new peer, gossipd hands up to lightningd, with global/local features
and the peer fd and a gossip fd using `gossip_peer_connected`
3. If lightningd doesn't want it, it just hands the peerfd and global/local
features back to gossipd using `gossipctl_handle_peer`
4. If a peer sends a non-gossip msg (eg `open_channel`) the gossipd sends
it up using `gossip_peer_nongossip`.
5. If lightningd wants to fund a channel, it simply calls `release_channel`.
Notes:
* There's no more "unique_id": we use the peer id.
* For the moment, we don't ask gossipd when we're told to list peers, so
connected peers without a channel don't appear in the JSON getpeers API.
* We add a `gossipctl_peer_addrhint` for the moment, so you can connect to
a specific ip/port, but using other sources is a TODO.
* We now (correctly) only give up on reaching a peer after we exchange init
messages, which changes the test_disconnect case.
Signed-off-by: Rusty Russell <rusty@rustcorp.com.au>
2017-10-11 12:09:49 +02:00
# Should have 3 connect fails.
for d in disconnects :
l1 . daemon . wait_for_log ( ' Failed connected out for {} , will try again '
. format ( l2 . info [ ' id ' ] ) )
2017-10-24 04:06:14 +02:00
@unittest.skipIf ( not DEVELOPER , " needs DEVELOPER=1 " )
2017-08-14 22:36:42 +02:00
def test_disconnect_funder ( self ) :
2017-05-25 04:15:57 +02:00
# Now error on funder side duringchannel open.
disconnects = [ ' -WIRE_OPEN_CHANNEL ' ,
' @WIRE_OPEN_CHANNEL ' ,
' +WIRE_OPEN_CHANNEL ' ,
' -WIRE_FUNDING_CREATED ' ,
' @WIRE_FUNDING_CREATED ' ]
2017-08-28 17:53:01 +02:00
l1 = self . node_factory . get_node ( disconnect = disconnects )
l2 = self . node_factory . get_node ( )
2017-05-25 04:15:57 +02:00
addr = l1 . rpc . newaddr ( ) [ ' address ' ]
txid = l1 . bitcoin . rpc . sendtoaddress ( addr , 20000 / 10 * * 6 )
tx = l1 . bitcoin . rpc . getrawtransaction ( txid )
l1 . rpc . addfunds ( tx )
for d in disconnects :
2017-12-04 10:12:22 +01:00
l1 . rpc . connect ( l2 . info [ ' id ' ] , ' localhost ' , l2 . info [ ' port ' ] )
2017-05-25 04:15:57 +02:00
self . assertRaises ( ValueError , l1 . rpc . fundchannel , l2 . info [ ' id ' ] , 20000 )
assert l1 . rpc . getpeer ( l2 . info [ ' id ' ] ) == None
2017-10-24 04:06:14 +02:00
@unittest.skipIf ( not DEVELOPER , " needs DEVELOPER=1 " )
2017-08-14 22:36:42 +02:00
def test_disconnect_fundee ( self ) :
2017-05-25 04:15:57 +02:00
# Now error on fundee side during channel open.
disconnects = [ ' -WIRE_ACCEPT_CHANNEL ' ,
' @WIRE_ACCEPT_CHANNEL ' ,
' +WIRE_ACCEPT_CHANNEL ' ]
2017-08-28 17:53:01 +02:00
l1 = self . node_factory . get_node ( )
l2 = self . node_factory . get_node ( disconnect = disconnects )
2017-05-25 04:15:57 +02:00
addr = l1 . rpc . newaddr ( ) [ ' address ' ]
txid = l1 . bitcoin . rpc . sendtoaddress ( addr , 20000 / 10 * * 6 )
tx = l1 . bitcoin . rpc . getrawtransaction ( txid )
l1 . rpc . addfunds ( tx )
for d in disconnects :
2017-12-04 10:12:22 +01:00
l1 . rpc . connect ( l2 . info [ ' id ' ] , ' localhost ' , l2 . info [ ' port ' ] )
2017-05-25 04:15:57 +02:00
self . assertRaises ( ValueError , l1 . rpc . fundchannel , l2 . info [ ' id ' ] , 20000 )
assert l1 . rpc . getpeer ( l2 . info [ ' id ' ] ) == None
2017-10-24 04:06:14 +02:00
@unittest.skipIf ( not DEVELOPER , " needs DEVELOPER=1 " )
2017-08-14 22:36:42 +02:00
def test_disconnect_half_signed ( self ) :
2017-05-25 04:15:57 +02:00
# Now, these are the corner cases. Fundee sends funding_signed,
# but funder doesn't receive it.
disconnects = [ ' @WIRE_FUNDING_SIGNED ' ]
2017-08-28 17:53:01 +02:00
l1 = self . node_factory . get_node ( )
l2 = self . node_factory . get_node ( disconnect = disconnects )
2017-05-25 04:15:57 +02:00
addr = l1 . rpc . newaddr ( ) [ ' address ' ]
txid = l1 . bitcoin . rpc . sendtoaddress ( addr , 20000 / 10 * * 6 )
tx = l1 . bitcoin . rpc . getrawtransaction ( txid )
l1 . rpc . addfunds ( tx )
2017-12-04 10:12:22 +01:00
l1 . rpc . connect ( l2 . info [ ' id ' ] , ' localhost ' , l2 . info [ ' port ' ] )
2017-05-25 04:15:57 +02:00
self . assertRaises ( ValueError , l1 . rpc . fundchannel , l2 . info [ ' id ' ] , 20000 )
# Fundee remembers, funder doesn't.
assert l1 . rpc . getpeer ( l2 . info [ ' id ' ] ) == None
assert l2 . rpc . getpeer ( l1 . info [ ' id ' ] ) [ ' peerid ' ] == l1 . info [ ' id ' ]
2017-10-24 04:06:14 +02:00
@unittest.skipIf ( not DEVELOPER , " needs DEVELOPER=1 " )
2017-06-20 08:11:03 +02:00
def test_reconnect_signed ( self ) :
2017-05-25 04:16:02 +02:00
# This will fail *after* both sides consider channel opening.
disconnects = [ ' +WIRE_FUNDING_SIGNED ' ]
2017-08-28 17:53:01 +02:00
l1 = self . node_factory . get_node ( )
l2 = self . node_factory . get_node ( disconnect = disconnects )
2017-05-25 04:16:02 +02:00
addr = l1 . rpc . newaddr ( ) [ ' address ' ]
txid = l1 . bitcoin . rpc . sendtoaddress ( addr , 20000 / 10 * * 6 )
tx = l1 . bitcoin . rpc . getrawtransaction ( txid )
l1 . rpc . addfunds ( tx )
2017-12-04 10:12:22 +01:00
l1 . rpc . connect ( l2 . info [ ' id ' ] , ' localhost ' , l2 . info [ ' port ' ] )
2017-05-25 04:16:02 +02:00
l1 . rpc . fundchannel ( l2 . info [ ' id ' ] , 20000 )
# They haven't forgotten each other.
assert l1 . rpc . getpeer ( l2 . info [ ' id ' ] ) [ ' peerid ' ] == l2 . info [ ' id ' ]
assert l2 . rpc . getpeer ( l1 . info [ ' id ' ] ) [ ' peerid ' ] == l1 . info [ ' id ' ]
# Technically, this is async to fundchannel.
l1 . daemon . wait_for_log ( ' sendrawtx exit 0 ' )
2017-07-01 03:41:17 +02:00
# Wait for reconnect, awaiting lockin..
2017-05-25 04:16:02 +02:00
l1 . daemon . wait_for_log ( ' Peer has reconnected, state CHANNELD_AWAITING_LOCKIN ' ) ;
2017-11-28 05:39:22 +01:00
l1 . bitcoin . generate_block ( 6 )
2017-05-25 04:16:02 +02:00
l1 . daemon . wait_for_log ( ' -> CHANNELD_NORMAL ' )
l2 . daemon . wait_for_log ( ' -> CHANNELD_NORMAL ' )
2017-08-14 22:36:42 +02:00
2017-10-24 04:06:14 +02:00
@unittest.skipIf ( not DEVELOPER , " needs DEVELOPER=1 " )
2017-09-06 04:40:04 +02:00
def test_reconnect_openingd ( self ) :
# Openingd thinks we're still opening; funder reconnects..
disconnects = [ ' 0WIRE_ACCEPT_CHANNEL ' ]
l1 = self . node_factory . get_node ( )
l2 = self . node_factory . get_node ( disconnect = disconnects )
2017-12-04 10:12:22 +01:00
l1 . rpc . connect ( l2 . info [ ' id ' ] , ' localhost ' , l2 . info [ ' port ' ] )
2017-09-06 04:40:04 +02:00
addr = l1 . rpc . newaddr ( ) [ ' address ' ]
txid = l1 . bitcoin . rpc . sendtoaddress ( addr , 20000 / 10 * * 6 )
tx = l1 . bitcoin . rpc . getrawtransaction ( txid )
l1 . rpc . addfunds ( tx )
2017-10-11 12:00:50 +02:00
# l2 closes on l1, l1 forgets.
2017-09-06 04:40:04 +02:00
self . assertRaises ( ValueError , l1 . rpc . fundchannel , l2 . info [ ' id ' ] , 20000 )
assert l1 . rpc . getpeer ( l2 . info [ ' id ' ] ) == None
# Reconnect.
2017-12-04 10:12:22 +01:00
l1 . rpc . connect ( l2 . info [ ' id ' ] , ' localhost ' , l2 . info [ ' port ' ] )
2017-09-06 04:40:04 +02:00
2017-10-12 08:35:03 +02:00
# We should get a message about reconnecting.
2017-10-11 12:00:50 +02:00
l2 . daemon . wait_for_log ( ' Peer has reconnected, state OPENINGD ' )
2017-09-06 04:40:04 +02:00
# Should work fine.
l1 . rpc . fundchannel ( l2 . info [ ' id ' ] , 20000 )
l1 . daemon . wait_for_log ( ' sendrawtx exit 0 ' )
2017-10-11 12:00:50 +02:00
# Just to be sure, second openingd hand over to channeld.
2017-10-12 08:35:03 +02:00
l2 . daemon . wait_for_log ( ' lightning_openingd.*REPLY WIRE_OPENING_FUNDEE_REPLY with 2 fds ' )
2017-09-06 04:40:04 +02:00
2017-10-24 04:06:14 +02:00
@unittest.skipIf ( not DEVELOPER , " needs DEVELOPER=1 " )
2017-06-20 08:11:03 +02:00
def test_reconnect_normal ( self ) :
# Should reconnect fine even if locked message gets lost.
disconnects = [ ' -WIRE_FUNDING_LOCKED ' ,
' @WIRE_FUNDING_LOCKED ' ,
' +WIRE_FUNDING_LOCKED ' ]
2017-08-28 17:53:01 +02:00
l1 = self . node_factory . get_node ( disconnect = disconnects )
l2 = self . node_factory . get_node ( )
2017-12-04 10:12:22 +01:00
ret = l1 . rpc . connect ( l2 . info [ ' id ' ] , ' localhost ' , l2 . info [ ' port ' ] )
2017-06-20 08:11:03 +02:00
self . fund_channel ( l1 , l2 , 10 * * 6 )
2017-06-20 08:21:03 +02:00
2017-10-24 04:06:14 +02:00
@unittest.skipIf ( not DEVELOPER , " needs DEVELOPER=1 " )
2017-10-11 12:08:50 +02:00
def test_reconnect_sender_add1 ( self ) :
2017-06-20 08:21:03 +02:00
# Fail after add is OK, will cause payment failure though.
2017-10-11 12:08:50 +02:00
disconnects = [ ' -WIRE_UPDATE_ADD_HTLC-nocommit ' ,
' +WIRE_UPDATE_ADD_HTLC-nocommit ' ,
' @WIRE_UPDATE_ADD_HTLC-nocommit ' ]
l1 = self . node_factory . get_node ( disconnect = disconnects )
l2 = self . node_factory . get_node ( )
2017-12-04 10:12:22 +01:00
ret = l1 . rpc . connect ( l2 . info [ ' id ' ] , ' localhost ' , l2 . info [ ' port ' ] )
2017-10-11 12:08:50 +02:00
self . fund_channel ( l1 , l2 , 10 * * 6 )
amt = 200000000
2017-10-26 05:06:19 +02:00
rhash = l2 . rpc . invoice ( amt , ' test_reconnect_sender_add1 ' , ' desc ' ) [ ' rhash ' ]
2017-10-11 12:08:50 +02:00
assert l2 . rpc . listinvoice ( ' test_reconnect_sender_add1 ' ) [ 0 ] [ ' complete ' ] == False
route = [ { ' msatoshi ' : amt , ' id ' : l2 . info [ ' id ' ] , ' delay ' : 5 , ' channel ' : ' 1:1:1 ' } ]
for i in range ( 0 , len ( disconnects ) ) :
self . assertRaises ( ValueError , l1 . rpc . sendpay , to_json ( route ) , rhash )
# Wait for reconnection.
l1 . daemon . wait_for_log ( ' Already have funding locked in ' )
# This will send commit, so will reconnect as required.
l1 . rpc . sendpay ( to_json ( route ) , rhash )
2017-10-24 04:06:14 +02:00
@unittest.skipIf ( not DEVELOPER , " needs DEVELOPER=1 " )
2017-10-11 12:08:50 +02:00
def test_reconnect_sender_add ( self ) :
disconnects = [ ' -WIRE_COMMITMENT_SIGNED ' ,
2017-06-20 08:23:03 +02:00
' @WIRE_COMMITMENT_SIGNED ' ,
' +WIRE_COMMITMENT_SIGNED ' ,
' -WIRE_REVOKE_AND_ACK ' ,
' @WIRE_REVOKE_AND_ACK ' ,
' +WIRE_REVOKE_AND_ACK ' ]
2017-08-28 17:53:01 +02:00
l1 = self . node_factory . get_node ( disconnect = disconnects )
l2 = self . node_factory . get_node ( )
2017-12-04 10:12:22 +01:00
ret = l1 . rpc . connect ( l2 . info [ ' id ' ] , ' localhost ' , l2 . info [ ' port ' ] )
2017-06-20 08:21:03 +02:00
self . fund_channel ( l1 , l2 , 10 * * 6 )
amt = 200000000
2017-10-26 05:06:19 +02:00
rhash = l2 . rpc . invoice ( amt , ' testpayment ' , ' desc ' ) [ ' rhash ' ]
2017-06-26 23:21:29 +02:00
assert l2 . rpc . listinvoice ( ' testpayment ' ) [ 0 ] [ ' complete ' ] == False
2017-06-20 08:21:03 +02:00
route = [ { ' msatoshi ' : amt , ' id ' : l2 . info [ ' id ' ] , ' delay ' : 5 , ' channel ' : ' 1:1:1 ' } ]
2017-06-26 23:21:29 +02:00
# This will send commit, so will reconnect as required.
2017-06-20 08:21:03 +02:00
l1 . rpc . sendpay ( to_json ( route ) , rhash )
2017-06-20 08:23:03 +02:00
# Should have printed this for every reconnect.
2017-10-11 12:08:50 +02:00
for i in range ( 0 , len ( disconnects ) ) :
2017-06-20 08:23:03 +02:00
l1 . daemon . wait_for_log ( ' Already have funding locked in ' )
2017-10-24 04:06:14 +02:00
@unittest.skipIf ( not DEVELOPER , " needs DEVELOPER=1 " )
2017-06-20 08:23:03 +02:00
def test_reconnect_receiver_add ( self ) :
disconnects = [ ' -WIRE_COMMITMENT_SIGNED ' ,
' @WIRE_COMMITMENT_SIGNED ' ,
' +WIRE_COMMITMENT_SIGNED ' ,
' -WIRE_REVOKE_AND_ACK ' ,
' @WIRE_REVOKE_AND_ACK ' ,
' +WIRE_REVOKE_AND_ACK ' ]
2017-08-28 17:53:01 +02:00
l1 = self . node_factory . get_node ( )
l2 = self . node_factory . get_node ( disconnect = disconnects )
2017-12-04 10:12:22 +01:00
ret = l1 . rpc . connect ( l2 . info [ ' id ' ] , ' localhost ' , l2 . info [ ' port ' ] )
2017-06-20 08:23:03 +02:00
self . fund_channel ( l1 , l2 , 10 * * 6 )
amt = 200000000
2017-10-26 05:06:19 +02:00
rhash = l2 . rpc . invoice ( amt , ' testpayment2 ' , ' desc ' ) [ ' rhash ' ]
2017-06-20 08:23:03 +02:00
assert l2 . rpc . listinvoice ( ' testpayment2 ' ) [ 0 ] [ ' complete ' ] == False
2017-06-20 08:21:03 +02:00
2017-06-20 08:23:03 +02:00
route = [ { ' msatoshi ' : amt , ' id ' : l2 . info [ ' id ' ] , ' delay ' : 5 , ' channel ' : ' 1:1:1 ' } ]
l1 . rpc . sendpay ( to_json ( route ) , rhash )
for i in range ( len ( disconnects ) ) :
l1 . daemon . wait_for_log ( ' Already have funding locked in ' )
assert l2 . rpc . listinvoice ( ' testpayment2 ' ) [ 0 ] [ ' complete ' ] == True
2017-10-24 04:06:14 +02:00
@unittest.skipIf ( not DEVELOPER , " needs DEVELOPER=1 " )
2017-06-20 08:23:03 +02:00
def test_reconnect_receiver_fulfill ( self ) :
2017-10-11 12:09:49 +02:00
# Ordering matters: after +WIRE_UPDATE_FULFILL_HTLC, channeld
# will continue and try to send WIRE_COMMITMENT_SIGNED: if
# that's the next failure, it will do two in one run.
disconnects = [ ' @WIRE_UPDATE_FULFILL_HTLC ' ,
2017-06-20 08:23:03 +02:00
' +WIRE_UPDATE_FULFILL_HTLC ' ,
2017-10-11 12:09:49 +02:00
' -WIRE_UPDATE_FULFILL_HTLC ' ,
2017-06-20 08:23:03 +02:00
' -WIRE_COMMITMENT_SIGNED ' ,
' @WIRE_COMMITMENT_SIGNED ' ,
' +WIRE_COMMITMENT_SIGNED ' ,
' -WIRE_REVOKE_AND_ACK ' ,
' @WIRE_REVOKE_AND_ACK ' ,
' +WIRE_REVOKE_AND_ACK ' ]
2017-08-28 17:53:01 +02:00
l1 = self . node_factory . get_node ( )
l2 = self . node_factory . get_node ( disconnect = disconnects )
2017-12-04 10:12:22 +01:00
ret = l1 . rpc . connect ( l2 . info [ ' id ' ] , ' localhost ' , l2 . info [ ' port ' ] )
2017-06-20 08:23:03 +02:00
self . fund_channel ( l1 , l2 , 10 * * 6 )
amt = 200000000
2017-10-26 05:06:19 +02:00
rhash = l2 . rpc . invoice ( amt , ' testpayment2 ' , ' desc ' ) [ ' rhash ' ]
2017-06-20 08:23:03 +02:00
assert l2 . rpc . listinvoice ( ' testpayment2 ' ) [ 0 ] [ ' complete ' ] == False
route = [ { ' msatoshi ' : amt , ' id ' : l2 . info [ ' id ' ] , ' delay ' : 5 , ' channel ' : ' 1:1:1 ' } ]
l1 . rpc . sendpay ( to_json ( route ) , rhash )
for i in range ( len ( disconnects ) ) :
l1 . daemon . wait_for_log ( ' Already have funding locked in ' )
assert l2 . rpc . listinvoice ( ' testpayment2 ' ) [ 0 ] [ ' complete ' ] == True
2017-07-05 07:33:40 +02:00
2017-10-24 04:06:14 +02:00
@unittest.skipIf ( not DEVELOPER , " needs DEVELOPER=1 " )
2017-07-05 07:33:40 +02:00
def test_shutdown_reconnect ( self ) :
disconnects = [ ' -WIRE_SHUTDOWN ' ,
' @WIRE_SHUTDOWN ' ,
' +WIRE_SHUTDOWN ' ]
2017-08-28 17:53:01 +02:00
l1 = self . node_factory . get_node ( disconnect = disconnects )
l2 = self . node_factory . get_node ( )
2017-12-04 10:12:22 +01:00
l1 . rpc . connect ( l2 . info [ ' id ' ] , ' localhost ' , l2 . info [ ' port ' ] )
2017-07-05 07:33:40 +02:00
self . fund_channel ( l1 , l2 , 10 * * 6 )
2017-07-05 07:33:40 +02:00
self . pay ( l1 , l2 , 200000000 )
2017-07-05 07:33:40 +02:00
assert l1 . bitcoin . rpc . getmempoolinfo ( ) [ ' size ' ] == 0
# This should return, then close.
l1 . rpc . close ( l2 . info [ ' id ' ] ) ;
l1 . daemon . wait_for_log ( ' -> CHANNELD_SHUTTING_DOWN ' )
l2 . daemon . wait_for_log ( ' -> CHANNELD_SHUTTING_DOWN ' )
l1 . daemon . wait_for_log ( ' -> CLOSINGD_SIGEXCHANGE ' )
l2 . daemon . wait_for_log ( ' -> CLOSINGD_SIGEXCHANGE ' )
2017-07-05 07:33:40 +02:00
# And should put closing into mempool (happens async, so
# CLOSINGD_COMPLETE may come first).
l1 . daemon . wait_for_logs ( [ ' sendrawtx exit 0 ' , ' -> CLOSINGD_COMPLETE ' ] )
l2 . daemon . wait_for_logs ( [ ' sendrawtx exit 0 ' , ' -> CLOSINGD_COMPLETE ' ] )
2017-07-05 07:33:40 +02:00
assert l1 . bitcoin . rpc . getmempoolinfo ( ) [ ' size ' ] == 1
2017-08-14 22:36:42 +02:00
2017-10-24 04:06:14 +02:00
@unittest.skipIf ( not DEVELOPER , " needs DEVELOPER=1 " )
2017-07-05 07:33:40 +02:00
def test_closing_negotiation_reconnect ( self ) :
disconnects = [ ' -WIRE_CLOSING_SIGNED ' ,
' @WIRE_CLOSING_SIGNED ' ,
' +WIRE_CLOSING_SIGNED ' ]
2017-08-28 17:53:01 +02:00
l1 = self . node_factory . get_node ( disconnect = disconnects )
l2 = self . node_factory . get_node ( )
2017-12-04 10:12:22 +01:00
l1 . rpc . connect ( l2 . info [ ' id ' ] , ' localhost ' , l2 . info [ ' port ' ] )
2017-07-05 07:33:40 +02:00
self . fund_channel ( l1 , l2 , 10 * * 6 )
2017-07-05 07:33:40 +02:00
self . pay ( l1 , l2 , 200000000 )
2017-07-05 07:33:40 +02:00
assert l1 . bitcoin . rpc . getmempoolinfo ( ) [ ' size ' ] == 0
# This should return, then close.
l1 . rpc . close ( l2 . info [ ' id ' ] ) ;
l1 . daemon . wait_for_log ( ' -> CHANNELD_SHUTTING_DOWN ' )
l2 . daemon . wait_for_log ( ' -> CHANNELD_SHUTTING_DOWN ' )
l1 . daemon . wait_for_log ( ' -> CLOSINGD_SIGEXCHANGE ' )
l2 . daemon . wait_for_log ( ' -> CLOSINGD_SIGEXCHANGE ' )
2017-07-05 07:33:40 +02:00
# And should put closing into mempool (happens async, so
# CLOSINGD_COMPLETE may come first).
l1 . daemon . wait_for_logs ( [ ' sendrawtx exit 0 ' , ' -> CLOSINGD_COMPLETE ' ] )
l2 . daemon . wait_for_logs ( [ ' sendrawtx exit 0 ' , ' -> CLOSINGD_COMPLETE ' ] )
2017-07-05 07:33:40 +02:00
assert l1 . bitcoin . rpc . getmempoolinfo ( ) [ ' size ' ] == 1
2017-07-05 07:33:40 +02:00
2017-05-31 13:15:03 +02:00
def test_json_addfunds ( self ) :
sat = 10 * * 6
2017-08-28 17:53:01 +02:00
l1 = self . node_factory . get_node ( )
2017-05-31 13:15:03 +02:00
addr = l1 . rpc . newaddr ( ) [ ' address ' ]
txid = l1 . bitcoin . rpc . sendtoaddress ( addr , 0.01 )
tx = l1 . bitcoin . rpc . getrawtransaction ( txid )
# The first time should succeed
assert l1 . rpc . addfunds ( tx ) == { " outputs " : 1 , " satoshis " : sat }
# Second time should fail, we already have those funds
self . assertRaises ( ValueError , l1 . rpc . addfunds , tx )
2017-06-20 16:41:49 +02:00
def test_withdraw ( self ) :
amount = 1000000
2017-08-28 17:53:01 +02:00
l1 = self . node_factory . get_node ( )
2017-06-20 16:41:49 +02:00
addr = l1 . rpc . newaddr ( ) [ ' address ' ]
# Add some funds to withdraw later
for i in range ( 10 ) :
txid = l1 . bitcoin . rpc . sendtoaddress ( addr , amount / 10 * * 8 + 0.01 )
tx = l1 . bitcoin . rpc . getrawtransaction ( txid )
l1 . rpc . addfunds ( tx )
# Reach around into the db to check that outputs were added
db = sqlite3 . connect ( os . path . join ( l1 . daemon . lightning_dir , " lightningd.sqlite3 " ) )
c = db . cursor ( )
c . execute ( ' SELECT COUNT(*) FROM outputs WHERE status=0 ' )
assert ( c . fetchone ( ) [ 0 ] == 10 )
waddr = l1 . bitcoin . rpc . getnewaddress ( )
# Now attempt to withdraw some (making sure we collect multiple inputs)
self . assertRaises ( ValueError , l1 . rpc . withdraw , ' not an address ' , amount )
self . assertRaises ( ValueError , l1 . rpc . withdraw , waddr , ' not an amount ' )
self . assertRaises ( ValueError , l1 . rpc . withdraw , waddr , - amount )
out = l1 . rpc . withdraw ( waddr , 2 * amount )
# Make sure bitcoind received the withdrawal
unspent = l1 . bitcoin . rpc . listunspent ( 0 )
withdrawal = [ u for u in unspent if u [ ' txid ' ] == out [ ' txid ' ] ]
assert ( len ( withdrawal ) == 1 )
assert ( withdrawal [ 0 ] [ ' amount ' ] == Decimal ( ' 0.02 ' ) )
# Now make sure two of them were marked as spent
c = db . cursor ( )
c . execute ( ' SELECT COUNT(*) FROM outputs WHERE status=2 ' )
assert ( c . fetchone ( ) [ 0 ] == 2 )
2017-09-01 14:40:52 +02:00
def test_funding_change ( self ) :
""" Add some funds, fund a channel, and make sure we remember the change
"""
l1 , l2 = self . connect ( )
addr = l1 . rpc . newaddr ( ) [ ' address ' ]
txid = l1 . bitcoin . rpc . sendtoaddress ( addr , 0.1 )
tx = l1 . bitcoin . rpc . getrawtransaction ( txid )
l1 . rpc . addfunds ( tx )
outputs = l1 . db_query ( ' SELECT value FROM outputs WHERE status=0; ' )
assert len ( outputs ) == 1 and outputs [ 0 ] [ ' value ' ] == 10000000
l1 . rpc . fundchannel ( l2 . info [ ' id ' ] , 1000000 )
outputs = { r [ ' status ' ] : r [ ' value ' ] for r in l1 . db_query (
' SELECT status, SUM(value) AS value FROM outputs GROUP BY status; ' ) }
# The 10m out is spent and we have a change output of 9m-fee
assert outputs [ 0 ] > 8990000
assert outputs [ 2 ] == 10000000
2017-11-25 13:31:49 +01:00
def test_addfunds_from_block ( self ) :
""" Send funds to the daemon without telling it explicitly
"""
2017-12-02 07:36:15 +01:00
# Previous runs with same bitcoind can leave funds!
l1 = self . node_factory . get_node ( random_hsm = True )
2017-11-25 13:31:49 +01:00
addr = l1 . rpc . newaddr ( ) [ ' address ' ]
txid = l1 . bitcoin . rpc . sendtoaddress ( addr , 0.1 )
l1 . bitcoin . rpc . generate ( 1 )
wait_for ( lambda : len ( l1 . rpc . listfunds ( ) [ ' outputs ' ] ) == 1 )
outputs = l1 . db_query ( ' SELECT value FROM outputs WHERE status=0; ' )
assert len ( outputs ) == 1 and outputs [ 0 ] [ ' value ' ] == 10000000
# Now the same, but create a conflict between addfunds and from block
txid = l1 . bitcoin . rpc . sendtoaddress ( addr , 0.1 )
tx = l1 . bitcoin . rpc . getrawtransaction ( txid )
l1 . rpc . addfunds ( tx )
l1 . bitcoin . rpc . generate ( 1 )
time . sleep ( 5 )
2017-10-24 04:06:14 +02:00
@unittest.skipIf ( not DEVELOPER , " needs DEVELOPER=1 " )
2017-08-14 16:52:19 +02:00
def test_channel_persistence ( self ) :
2017-09-30 22:41:53 +02:00
# Start two nodes and open a channel (to remember). l2 will
# mysteriously die while committing the first HTLC so we can
# check that HTLCs reloaded from the DB work.
l1 = self . node_factory . get_node ( )
2017-10-11 12:08:50 +02:00
l2 = self . node_factory . get_node ( disconnect = [ ' =WIRE_COMMITMENT_SIGNED-nocommit ' ] )
2017-12-04 10:12:22 +01:00
l1 . rpc . connect ( l2 . info [ ' id ' ] , ' localhost ' , l2 . info [ ' port ' ] )
2017-08-14 16:52:19 +02:00
# Neither node should have a channel open, they are just connected
for n in ( l1 , l2 ) :
assert ( n . db_query ( ' SELECT COUNT(id) as count FROM channels; ' ) [ 0 ] [ ' count ' ] == 0 )
self . fund_channel ( l1 , l2 , 100000 )
peers = l1 . rpc . getpeers ( ) [ ' peers ' ]
assert ( len ( peers ) == 1 and peers [ 0 ] [ ' state ' ] == ' CHANNELD_NORMAL ' )
# Both nodes should now have exactly one channel in the database
for n in ( l1 , l2 ) :
assert ( n . db_query ( ' SELECT COUNT(id) as count FROM channels; ' ) [ 0 ] [ ' count ' ] == 1 )
2017-09-30 22:41:53 +02:00
# Fire off a sendpay request, it'll get interrupted by a restart
fut = self . executor . submit ( self . pay , l1 , l2 , 10000 )
# Wait for it to be committed to, i.e., stored in the DB
l1 . daemon . wait_for_log ( ' peer_in WIRE_COMMITMENT_SIGNED ' )
2017-08-24 14:15:30 +02:00
# Stop l2, l1 will reattempt to connect
2017-09-30 22:41:53 +02:00
print ( " Killing l2 in mid HTLC " )
l2 . daemon . proc . terminate ( )
# Clear the disconnect and timer stop so we can proceed normally
l2 . daemon . cmd_line = [ e for e in l2 . daemon . cmd_line if ' disconnect ' not in e ]
print ( " " . join ( l2 . daemon . cmd_line + [ ' --dev-debugger=channeld ' ] ) )
2017-08-14 16:52:19 +02:00
2017-08-24 14:15:30 +02:00
# Wait for l1 to notice
2017-08-17 15:30:24 +02:00
wait_for ( lambda : not l1 . rpc . getpeers ( ) [ ' peers ' ] [ 0 ] [ ' connected ' ] )
2017-08-14 16:52:19 +02:00
# Now restart l1 and it should reload peers/channels from the DB
2017-08-17 15:30:24 +02:00
l2 . daemon . start ( )
wait_for ( lambda : len ( l2 . rpc . getpeers ( ) [ ' peers ' ] ) == 1 )
2017-08-14 16:52:19 +02:00
2017-09-30 22:41:53 +02:00
# Wait for the restored HTLC to finish
wait_for ( lambda : l1 . rpc . getpeers ( ) [ ' peers ' ] [ 0 ] [ ' msatoshi_to_us ' ] == 99990000 , interval = 1 )
2017-08-17 15:30:24 +02:00
wait_for ( lambda : len ( [ p for p in l1 . rpc . getpeers ( ) [ ' peers ' ] if p [ ' connected ' ] ] ) , interval = 1 )
wait_for ( lambda : len ( [ p for p in l2 . rpc . getpeers ( ) [ ' peers ' ] if p [ ' connected ' ] ] ) , interval = 1 )
2017-08-14 16:52:19 +02:00
2017-08-17 15:30:24 +02:00
# Now make sure this is really functional by sending a payment
self . pay ( l1 , l2 , 10000 )
2017-08-24 14:15:30 +02:00
time . sleep ( 1 )
assert l1 . rpc . getpeers ( ) [ ' peers ' ] [ 0 ] [ ' msatoshi_to_us ' ] == 99980000
assert l2 . rpc . getpeers ( ) [ ' peers ' ] [ 0 ] [ ' msatoshi_to_us ' ] == 20000
# Finally restart l1, and make sure it remembers
2017-09-28 05:31:47 +02:00
l1 . stop ( )
2017-08-24 14:15:30 +02:00
l1 . daemon . start ( )
assert l1 . rpc . getpeers ( ) [ ' peers ' ] [ 0 ] [ ' msatoshi_to_us ' ] == 99980000
2017-08-14 16:52:19 +02:00
2017-09-14 16:09:45 +02:00
def test_gossip_badsig ( self ) :
l1 = self . node_factory . get_node ( )
l2 = self . node_factory . get_node ( )
l3 = self . node_factory . get_node ( )
# l2 connects to both, so l1 can't reconnect and thus l2 drops to chain
2017-12-04 10:12:22 +01:00
l2 . rpc . connect ( l1 . info [ ' id ' ] , ' localhost ' , l1 . info [ ' port ' ] )
l2 . rpc . connect ( l3 . info [ ' id ' ] , ' localhost ' , l3 . info [ ' port ' ] )
2017-09-14 16:09:45 +02:00
self . fund_channel ( l2 , l1 , 10 * * 6 )
self . fund_channel ( l2 , l3 , 10 * * 6 )
# Wait for route propagation.
2017-11-28 05:39:22 +01:00
l1 . bitcoin . generate_block ( 5 )
2017-09-14 16:09:45 +02:00
l1 . daemon . wait_for_log ( ' Received node_announcement for node {} '
. format ( l3 . info [ ' id ' ] ) )
assert not l1 . daemon . is_in_log ( ' signature verification failed ' )
assert not l2 . daemon . is_in_log ( ' signature verification failed ' )
assert not l3 . daemon . is_in_log ( ' signature verification failed ' )
2017-11-13 16:15:10 +01:00
def test_waitanyinvoice ( self ) :
""" Test various variants of waiting for the next invoice to complete.
"""
l1 , l2 = self . connect ( )
inv1 = l2 . rpc . invoice ( 1000 , ' inv1 ' , ' inv1 ' )
inv2 = l2 . rpc . invoice ( 1000 , ' inv2 ' , ' inv2 ' )
inv3 = l2 . rpc . invoice ( 1000 , ' inv3 ' , ' inv3 ' )
print ( " AAAA " , l2 . rpc . listinvoice ( ) )
self . fund_channel ( l1 , l2 , 10 * * 6 )
2017-11-28 05:39:22 +01:00
l1 . bitcoin . generate_block ( 6 )
2017-11-13 16:15:10 +01:00
l1 . daemon . wait_for_log ( ' Received node_announcement for node {} ' . format ( l2 . info [ ' id ' ] ) )
# Attempt to wait for the first invoice
f = self . executor . submit ( l2 . rpc . waitanyinvoice )
time . sleep ( 1 )
# The call to waitanyinvoice should not have returned just yet
assert not f . done ( )
# Now pay the first two invoices and make sure we notice
l1 . rpc . pay ( inv1 [ ' bolt11 ' ] )
l1 . rpc . pay ( inv2 [ ' bolt11 ' ] )
r = f . result ( timeout = 5 )
assert r [ ' label ' ] == ' inv1 '
# This one should return immediately with inv2
r = self . executor . submit ( l2 . rpc . waitanyinvoice , ' inv1 ' ) . result ( timeout = 5 )
assert r [ ' label ' ] == ' inv2 '
# Now spawn the next waiter
f = self . executor . submit ( l2 . rpc . waitanyinvoice , ' inv2 ' )
time . sleep ( 1 )
assert not f . done ( )
l1 . rpc . pay ( inv3 [ ' bolt11 ' ] )
r = f . result ( timeout = 5 )
assert r [ ' label ' ] == ' inv3 '
self . assertRaises ( ValueError , l2 . rpc . waitanyinvoice , ' doesntexist ' )
2017-11-17 15:09:38 +01:00
def test_channel_reenable ( self ) :
l1 , l2 = self . line_graph ( n = 2 )
2017-11-28 05:39:22 +01:00
l1 . bitcoin . generate_block ( 6 )
2017-11-17 15:09:38 +01:00
l1 . daemon . wait_for_log ( ' Received node_announcement for node {} ' . format ( l2 . info [ ' id ' ] ) )
l2 . daemon . wait_for_log ( ' Received node_announcement for node {} ' . format ( l1 . info [ ' id ' ] ) )
# Both directions should be active before the restart
assert [ c [ ' active ' ] for c in l1 . rpc . getchannels ( ) [ ' channels ' ] ] == [ True , True ]
# Restart l2, will cause l1 to reconnect
l2 . stop ( )
l2 . daemon . start ( )
# Now they should sync and re-establish again
l1 . daemon . wait_for_log ( ' Received node_announcement for node {} ' . format ( l2 . info [ ' id ' ] ) )
l2 . daemon . wait_for_log ( ' Received node_announcement for node {} ' . format ( l1 . info [ ' id ' ] ) )
assert [ c [ ' active ' ] for c in l1 . rpc . getchannels ( ) [ ' channels ' ] ] == [ True , True ]
2017-11-21 06:26:59 +01:00
@unittest.skipIf ( not DEVELOPER , " needs DEVELOPER=1 " )
def test_update_fee ( self ) :
l1 , l2 = self . connect ( )
chanid = self . fund_channel ( l1 , l2 , 10 * * 6 )
# Make l1 send out feechange.
l1 . rpc . dev_setfees ( ' 14000 ' )
l2 . daemon . wait_for_log ( ' peer updated fee to 14000 ' )
# Now make sure an HTLC works.
# (First wait for route propagation.)
2017-11-28 05:39:22 +01:00
bitcoind . generate_block ( 6 )
2017-11-21 06:26:59 +01:00
l1 . daemon . wait_for_logs ( [ ' Received channel_update for channel {} \ (0 \ ) '
. format ( chanid ) ,
' Received channel_update for channel {} \ (1 \ ) '
. format ( chanid ) ] )
# Make payments.
self . pay ( l1 , l2 , 200000000 )
self . pay ( l2 , l1 , 100000000 )
# Now shutdown cleanly.
l1 . rpc . close ( l2 . info [ ' id ' ] ) ;
l1 . daemon . wait_for_log ( ' -> CLOSINGD_COMPLETE ' )
l2 . daemon . wait_for_log ( ' -> CLOSINGD_COMPLETE ' )
# And should put closing into mempool.
l1 . daemon . wait_for_log ( ' sendrawtx exit 0 ' )
l2 . daemon . wait_for_log ( ' sendrawtx exit 0 ' )
2017-11-28 05:39:22 +01:00
bitcoind . generate_block ( 1 )
2017-11-21 06:26:59 +01:00
l1 . daemon . wait_for_log ( ' -> ONCHAIND_MUTUAL ' )
l2 . daemon . wait_for_log ( ' -> ONCHAIND_MUTUAL ' )
2017-11-28 05:39:22 +01:00
bitcoind . generate_block ( 99 )
2017-11-21 06:26:59 +01:00
l1 . daemon . wait_for_log ( ' onchaind complete, forgetting peer ' )
l2 . daemon . wait_for_log ( ' onchaind complete, forgetting peer ' )
@unittest.skipIf ( not DEVELOPER , " needs DEVELOPER=1 " )
def test_update_fee_reconnect ( self ) :
# Disconnect after first commitsig.
disconnects = [ ' +WIRE_COMMITMENT_SIGNED ' ]
l1 = self . node_factory . get_node ( disconnect = disconnects )
l2 = self . node_factory . get_node ( )
2017-12-04 10:12:22 +01:00
l1 . rpc . connect ( l2 . info [ ' id ' ] , ' localhost ' , l2 . info [ ' port ' ] )
2017-11-21 06:26:59 +01:00
self . fund_channel ( l1 , l2 , 10 * * 6 )
# Make l1 send out feechange; triggers disconnect/reconnect.
l1 . rpc . dev_setfees ( ' 14000 ' )
l1 . daemon . wait_for_log ( ' Setting REMOTE feerate to 14000 ' )
l2 . daemon . wait_for_log ( ' Setting LOCAL feerate to 14000 ' )
l1 . daemon . wait_for_log ( ' dev_disconnect: \ +WIRE_COMMITMENT_SIGNED ' )
# Wait for reconnect....
l1 . daemon . wait_for_log ( ' Applying feerate 14000 to LOCAL ' )
self . pay ( l1 , l2 , 200000000 )
self . pay ( l2 , l1 , 100000000 )
# They should both have gotten commits with correct feerate.
assert l1 . daemon . is_in_log ( ' got commitsig [0-9]*: feerate 14000 ' )
assert l2 . daemon . is_in_log ( ' got commitsig [0-9]*: feerate 14000 ' )
# Now shutdown cleanly.
l1 . rpc . close ( l2 . info [ ' id ' ] ) ;
l1 . daemon . wait_for_log ( ' -> CLOSINGD_COMPLETE ' )
l2 . daemon . wait_for_log ( ' -> CLOSINGD_COMPLETE ' )
# And should put closing into mempool.
l1 . daemon . wait_for_log ( ' sendrawtx exit 0 ' )
l2 . daemon . wait_for_log ( ' sendrawtx exit 0 ' )
2017-11-28 05:39:22 +01:00
bitcoind . generate_block ( 1 )
2017-11-21 06:26:59 +01:00
l1 . daemon . wait_for_log ( ' -> ONCHAIND_MUTUAL ' )
l2 . daemon . wait_for_log ( ' -> ONCHAIND_MUTUAL ' )
2017-11-28 05:39:22 +01:00
bitcoind . generate_block ( 99 )
2017-11-21 06:26:59 +01:00
l1 . daemon . wait_for_log ( ' onchaind complete, forgetting peer ' )
l2 . daemon . wait_for_log ( ' onchaind complete, forgetting peer ' )
2017-11-23 13:08:36 +01:00
2017-01-14 20:32:32 +01:00
if __name__ == ' __main__ ' :
unittest . main ( verbosity = 2 )