2020-05-26 22:43:21 +02:00
from decimal import Decimal
2019-02-21 04:52:38 +01:00
from fixtures import * # noqa: F401,F403
2019-05-12 21:26:26 +02:00
from fixtures import TEST_NETWORK
2019-12-18 20:00:13 +01:00
from pyln . client import RpcError
2022-07-20 04:28:25 +02:00
from utils import wait_for , sync_blockheight , COMPAT , VALGRIND , DEVELOPER , TIMEOUT , only_one
2020-05-26 22:43:21 +02:00
import base64
2019-09-12 22:49:42 +02:00
import os
2019-12-18 20:00:13 +01:00
import pytest
2021-10-27 11:49:26 +02:00
import shutil
2019-12-18 20:00:13 +01:00
import time
2019-09-06 08:40:28 +02:00
import unittest
2019-08-05 13:52:29 +02:00
2019-05-12 21:26:26 +02:00
@unittest.skipIf ( TEST_NETWORK != ' regtest ' , " The DB migration is network specific due to the chain var. " )
2019-11-21 05:13:37 +01:00
def test_db_dangling_peer_fix ( node_factory , bitcoind ) :
# Make sure bitcoind doesn't think it's going backwards
bitcoind . generate_block ( 104 )
2019-02-21 04:52:38 +01:00
# This was taken from test_fail_unconfirmed() node.
l1 = node_factory . get_node ( dbfile = ' dangling-peer.sqlite3.xz ' )
l2 = node_factory . get_node ( )
# Must match entry in db
assert l2 . info [ ' id ' ] == ' 022d223620a359a47ff7f7ac447c85c46c923da53389221a0054c11c1e3ca31d59 '
# This time it should work! (Connect *in* since l1 thinks it has UTXOs
# it doesn't have).
l1 . rpc . connect ( l2 . info [ ' id ' ] , ' localhost ' , l2 . port )
# Make sure l2 has register connection
l2 . daemon . wait_for_log ( ' Handed peer, entering loop ' )
2020-09-24 14:06:36 +02:00
l2 . fundchannel ( l1 , 200000 , wait_for_active = True )
2019-08-05 13:52:29 +02:00
2019-09-18 21:55:41 +02:00
@unittest.skipIf ( TEST_NETWORK != ' regtest ' , " Address is network specific " )
def test_block_backfill ( node_factory , bitcoind , chainparams ) :
2019-08-05 13:52:29 +02:00
""" Test whether we backfill data from the blockchain correctly.
For normal operation we will process any block after the initial start
height , or rescan height , but for gossip we actually also need to backfill
the blocks we skipped initially . We do so on - demand , whenever we see a
channel_announcement referencing a blockheight we haven ' t processed yet,
we fetch the entire block , extract P2WSH outputs and ask ` bitcoin
gettxout ` for each of them . We then store the block header in the ` blocks `
table and the unspent outputs in the ` utxoset ` table .
The test consist of two nodes opening a channel at height X , and an
unrelated P2WSH transaction being sent at the same height ( will be used to
check for completeness of the backfill ) . Then a second node starts at
height X + 100 and connect to one of the nodes . It should not have the block
in its DB before connecting . After connecting it should sync the gossip ,
triggering a backfill of block X , and all associated P2WSH outputs .
"""
# Need to manually open the channels later since otherwise we can't have a
# tx in the same block (`line_graph` with `fundchannel=True` generates
# blocks).
l1 , l2 = node_factory . line_graph ( 2 , fundchannel = False )
# Get some funds to l1
addr = l1 . rpc . newaddr ( ) [ ' bech32 ' ]
bitcoind . rpc . sendtoaddress ( addr , 1 )
bitcoind . generate_block ( 1 )
2019-08-07 12:01:18 +02:00
wait_for ( lambda : len ( l1 . rpc . listfunds ( ) [ ' outputs ' ] ) == 1 )
2019-08-05 13:52:29 +02:00
# Now send the needle we will go looking for later:
2019-08-07 12:01:18 +02:00
bitcoind . rpc . sendtoaddress ( ' bcrt1qtwxd8wg5eanumk86vfeujvp48hfkgannf77evggzct048wggsrxsum2pmm ' , 0.00031337 )
2019-08-05 13:52:29 +02:00
l1 . rpc . fundchannel ( l2 . info [ ' id ' ] , 10 * * 6 , announce = True )
wait_for ( lambda : len ( bitcoind . rpc . getrawmempool ( ) ) == 2 )
# Confirm and get some distance between the funding and the l3 wallet birth date
bitcoind . generate_block ( 100 )
wait_for ( lambda : len ( l1 . rpc . listnodes ( ) [ ' nodes ' ] ) == 2 )
# Start the tester node, and connect it to l1. l0 should sync the gossip
# and call out to `bitcoind` to backfill the block.
l3 = node_factory . get_node ( )
heights = [ r [ ' height ' ] for r in l3 . db_query ( " SELECT height FROM blocks " ) ]
assert ( 103 not in heights )
l3 . rpc . connect ( l1 . info [ ' id ' ] , ' localhost ' , l1 . port )
2019-08-07 12:01:18 +02:00
# Make sure we have backfilled the block
2019-08-05 13:52:29 +02:00
wait_for ( lambda : len ( l3 . rpc . listnodes ( ) [ ' nodes ' ] ) == 2 )
heights = [ r [ ' height ' ] for r in l3 . db_query ( " SELECT height FROM blocks " ) ]
assert ( 103 in heights )
2019-08-07 12:01:18 +02:00
# Make sure we also have the needle we added to the haystack above
assert ( 31337 in [ r [ ' satoshis ' ] for r in l3 . db_query ( " SELECT satoshis FROM utxoset " ) ] )
2019-10-31 05:09:11 +01:00
# Make sure that l3 doesn't ask for more gossip and get a reply about
# the closed channel (hence Bad gossip msgs in log).
l3 . daemon . wait_for_log ( ' seeker: state = NORMAL ' )
2019-08-07 12:01:18 +02:00
# Now close the channel and make sure `l3` cleans up correctly:
2019-08-22 06:39:44 +02:00
txid = l1 . rpc . close ( l2 . info [ ' id ' ] ) [ ' txid ' ]
bitcoind . generate_block ( 1 , wait_for_mempool = txid )
2019-08-07 12:01:18 +02:00
wait_for ( lambda : len ( l3 . rpc . listchannels ( ) [ ' channels ' ] ) == 0 )
2019-08-09 18:01:31 +02:00
# Test that the max-channel-id is set correctly between
# restarts (with forgotten channel)
def test_max_channel_id ( node_factory , bitcoind ) :
# Create a channel between two peers.
# Close the channel and have 100 blocks happen (forget channel)
# Restart node, create channel again. Should succeed.
l1 , l2 = node_factory . line_graph ( 2 , fundchannel = True , wait_for_announce = True )
sync_blockheight ( bitcoind , [ l1 , l2 ] )
# Now shutdown cleanly.
l1 . rpc . close ( l2 . info [ ' id ' ] , 0 )
l1 . daemon . wait_for_log ( ' to CLOSINGD_COMPLETE ' )
l2 . daemon . wait_for_log ( ' to CLOSINGD_COMPLETE ' )
# And should put closing into mempool.
l1 . wait_for_channel_onchain ( l2 . info [ ' id ' ] )
l2 . wait_for_channel_onchain ( l1 . info [ ' id ' ] )
bitcoind . generate_block ( 101 )
2022-07-18 14:12:28 +02:00
wait_for ( lambda : only_one ( l1 . rpc . listpeers ( ) [ ' peers ' ] ) [ ' channels ' ] == [ ] )
wait_for ( lambda : only_one ( l2 . rpc . listpeers ( ) [ ' peers ' ] ) [ ' channels ' ] == [ ] )
2019-08-09 18:01:31 +02:00
# Stop l2, and restart
l2 . stop ( )
l2 . start ( )
# Reconnect
l1 . rpc . connect ( l2 . info [ ' id ' ] , ' localhost ' , l2 . port )
# Fundchannel again, should succeed.
l1 . rpc . fundchannel ( l2 . info [ ' id ' ] , 10 * * 5 )
2019-09-06 08:40:28 +02:00
@unittest.skipIf ( not COMPAT , " needs COMPAT to convert obsolete db " )
2019-09-12 22:49:42 +02:00
@unittest.skipIf ( os . getenv ( ' TEST_DB_PROVIDER ' , ' sqlite3 ' ) != ' sqlite3 ' , " This test is based on a sqlite3 snapshot " )
2019-09-18 21:55:41 +02:00
@unittest.skipIf ( TEST_NETWORK != ' regtest ' , " The network must match the DB snapshot " )
2019-11-21 05:13:37 +01:00
def test_scid_upgrade ( node_factory , bitcoind ) :
bitcoind . generate_block ( 1 )
2019-09-06 08:40:28 +02:00
# Created through the power of sed "s/X'\([0-9]*\)78\([0-9]*\)78\([0-9]*\)'/X'\13A\23A\3'/"
l1 = node_factory . get_node ( dbfile = ' oldstyle-scids.sqlite3.xz ' )
assert l1 . db_query ( ' SELECT short_channel_id from channels; ' ) == [ { ' short_channel_id ' : ' 103x1x1 ' } ]
assert l1 . db_query ( ' SELECT failchannel from payments; ' ) == [ { ' failchannel ' : ' 103x1x1 ' } ]
2019-12-18 20:00:13 +01:00
2021-05-20 02:15:12 +02:00
@unittest.skipIf ( not COMPAT , " needs COMPAT to convert obsolete db " )
@unittest.skipIf ( os . getenv ( ' TEST_DB_PROVIDER ' , ' sqlite3 ' ) != ' sqlite3 ' , " This test is based on a sqlite3 snapshot " )
@unittest.skipIf ( TEST_NETWORK != ' regtest ' , " The network must match the DB snapshot " )
def test_last_tx_inflight_psbt_upgrade ( node_factory , bitcoind ) :
bitcoind . generate_block ( 12 )
prior_txs = [ ' 02000000019CCCA2E59D863B00B5BD835BF7BA93CC257932D2C7CDBE51EFE2EE4A9D29DFCB01000000009DB0E280024A01000000000000220020BE7935A77CA9AB70A4B8B1906825637767FED3C00824AA90C988983587D68488F0820100000000002200209F4684DDB28ACDC73959BC194D1A25DF906F61ED030F52D163E6F1E247D32CBB9A3ED620 ' , ' 020000000122F9EBE38F54208545B681AD7F73A7AE3504A09C8201F502673D34E28424687C01000000009DB0E280024A01000000000000220020BE7935A77CA9AB70A4B8B1906825637767FED3C00824AA90C988983587D68488F0820100000000002200209F4684DDB28ACDC73959BC194D1A25DF906F61ED030F52D163E6F1E247D32CBB9A3ED620 ' ]
l1 = node_factory . get_node ( dbfile = ' upgrade_inflight.sqlite3.xz ' )
b64_last_txs = [ base64 . b64encode ( x [ ' last_tx ' ] ) . decode ( ' utf-8 ' ) for x in l1 . db_query ( ' SELECT last_tx FROM channel_funding_inflights ORDER BY channel_id, funding_feerate; ' ) ]
for i in range ( len ( b64_last_txs ) ) :
bpsbt = b64_last_txs [ i ]
psbt = bitcoind . rpc . decodepsbt ( bpsbt )
tx = prior_txs [ i ]
assert psbt [ ' tx ' ] [ ' txid ' ] == bitcoind . rpc . decoderawtransaction ( tx ) [ ' txid ' ]
funding_input = only_one ( psbt [ ' inputs ' ] )
assert funding_input [ ' witness_utxo ' ] [ ' amount ' ] == Decimal ( ' 0.001 ' )
assert funding_input [ ' witness_utxo ' ] [ ' scriptPubKey ' ] [ ' type ' ] == ' witness_v0_scripthash '
assert funding_input [ ' witness_script ' ] [ ' type ' ] == ' multisig '
2020-05-26 22:43:21 +02:00
@unittest.skipIf ( not COMPAT , " needs COMPAT to convert obsolete db " )
@unittest.skipIf ( os . getenv ( ' TEST_DB_PROVIDER ' , ' sqlite3 ' ) != ' sqlite3 ' , " This test is based on a sqlite3 snapshot " )
@unittest.skipIf ( TEST_NETWORK != ' regtest ' , " The network must match the DB snapshot " )
def test_last_tx_psbt_upgrade ( node_factory , bitcoind ) :
bitcoind . generate_block ( 12 )
prior_txs = [ ' 02000000018DD699861B00061E50937A233DB584BF8ED4C0BF50B44C0411F71B031A06455000000000000EF7A9800350C300000000000022002073356CFF7E1588F14935EF138E142ABEFB5F7E3D51DE942758DCD5A179449B6250A90600000000002200202DF545EA882889846C52FC5E111AC07CE07E0C09418AC15743A6F6284C2A4FA720A1070000000000160014E89954FAC8F7A2DCE51E095D7BEB5271C3F7DA56EF81DC20 ' , ' 02000000018A0AE4C63BCDF9D78B07EB4501BB23404FDDBC73973C592793F047BE1495074B010000000074D99980010A2D0F00000000002200203B8CB644781CBECA96BE8B2BF1827AFD908B3CFB5569AC74DAB9395E8DDA39E4C9555420 ' , ' 020000000135DAB2996E57762E3EC158C0D57D39F43CA657E882D93FC24F5FEBAA8F36ED9A0100000000566D1D800350C30000000000002200205679A7D06E1BD276AA25F56E9E4DF7E07D9837EFB0C5F63604F10CD9F766A03ED4DD0600000000001600147E5B5C8F4FC1A9484E259F92CA4CBB7FA2814EA49A6C070000000000220020AB6226DEBFFEFF4A741C01367FA3C875172483CFB3E327D0F8C7AA4C51EDECAA27AA4720 ' ]
l1 = node_factory . get_node ( dbfile = ' last_tx_upgrade.sqlite3.xz ' )
b64_last_txs = [ base64 . b64encode ( x [ ' last_tx ' ] ) . decode ( ' utf-8 ' ) for x in l1 . db_query ( ' SELECT last_tx FROM channels ORDER BY id; ' ) ]
for i in range ( len ( b64_last_txs ) ) :
bpsbt = b64_last_txs [ i ]
psbt = bitcoind . rpc . decodepsbt ( bpsbt )
tx = prior_txs [ i ]
assert psbt [ ' tx ' ] [ ' txid ' ] == bitcoind . rpc . decoderawtransaction ( tx ) [ ' txid ' ]
funding_input = only_one ( psbt [ ' inputs ' ] )
# Every opened channel was funded with the same amount: 1M sats
assert funding_input [ ' witness_utxo ' ] [ ' amount ' ] == Decimal ( ' 0.01 ' )
assert funding_input [ ' witness_utxo ' ] [ ' scriptPubKey ' ] [ ' type ' ] == ' witness_v0_scripthash '
assert funding_input [ ' witness_script ' ] [ ' type ' ] == ' multisig '
2020-06-17 01:25:32 +02:00
l1 . stop ( )
# Test again, but this time with a database with a closed channel + forgotten peer
# We need to get to block #232 from block #113
bitcoind . generate_block ( 232 - 113 )
# We need to give it a chance to update
time . sleep ( 2 )
l2 = node_factory . get_node ( dbfile = ' last_tx_closed.sqlite3.xz ' )
last_txs = [ x [ ' last_tx ' ] for x in l2 . db_query ( ' SELECT last_tx FROM channels ORDER BY id; ' ) ]
# The first tx should be psbt, the second should still be hex
bitcoind . rpc . decodepsbt ( base64 . b64encode ( last_txs [ 0 ] ) . decode ( ' utf-8 ' ) )
bitcoind . rpc . decoderawtransaction ( last_txs [ 1 ] . hex ( ) )
2020-05-26 22:43:21 +02:00
2020-07-29 04:03:03 +02:00
@unittest.skipIf ( os . getenv ( ' TEST_DB_PROVIDER ' , ' sqlite3 ' ) != ' sqlite3 ' , " This test is based on a sqlite3 snapshot " )
@unittest.skipIf ( TEST_NETWORK != ' regtest ' , " The network must match the DB snapshot " )
def test_backfill_scriptpubkeys ( node_factory , bitcoind ) :
bitcoind . generate_block ( 214 )
script_map = [
{
" txid " : " 2513F3340D493489811EAB440AC05650B5BC06290358972EB6A55533A9EED96A " ,
" scriptpubkey " : " 001438C10854C11E10CB3786460143C963C8530DF891 " ,
} , {
" txid " : " E380E18B6E810A464634B3A94B95AAA06B36A8982FD9D9D294982726EDC77DD3 " ,
" scriptpubkey " : " 001407DB91DA65EF06B385F4EA20BA05FAF286165C0B " ,
} , {
" txid " : " E9AE7C9A346F9B9E35868176F311F3F2EE5DB8B94A065963E26954E119C49A79 " ,
" scriptpubkey " : " 00147E5B5C8F4FC1A9484E259F92CA4CBB7FA2814EA4 " ,
} , {
" txid " : " 4C88F50BF00518E4FE3434ACA42351D5AC5FEEE17C35595DFBC3D1F4279F6EC1 " ,
" scriptpubkey " : " 0014D0EAC62FDCEE2D1881259BE9CDA4C43DE9050DB8 " ,
} , {
" txid " : " 55265C3CAFE98C355FE0A440DCC005CF5C3145280EAD44D6B903A45D2DF3619C " ,
" scriptpubkey " : " 0014D0EAC62FDCEE2D1881259BE9CDA4C43DE9050DB8 " ,
} , {
" txid " : " 06F6D1D29B175146381EAB59924EC438572D18A3701F8E4FDF4EE17DE78D31E3 " ,
" scriptpubkey " : " A9149551336F1E360F5AFB977F24CE72C744A82463D187 " ,
} , {
" txid " : " 91BCEC7867F3F97F4F575D1D9DEDF5CF22BDDE643B36C2D9E6097048334EE32A " ,
" scriptpubkey " : " 0014DFA9D65F06088E922A661C29797EE616F793C863 " ,
} ,
]
# Test the first time, all entries are with option_static_remotekey
l1 = node_factory . get_node ( node_id = 3 , dbfile = ' pubkey_regen.sqlite.xz ' )
results = l1 . db_query ( ' SELECT hex(prev_out_tx) AS txid, hex(scriptpubkey) AS script FROM outputs ' )
scripts = [ { ' txid ' : x [ ' txid ' ] , ' scriptpubkey ' : x [ ' script ' ] } for x in results ]
for exp , actual in zip ( script_map , scripts ) :
assert exp == actual
# Test again, without option_static_remotekey
script_map_2 = [
{
" txid " : " FF89677793AC6F39E4AEB9D393B45F1E3D902CBFA26B521C5C438345A6D36E54 " ,
" scriptpubkey " : " 001438C10854C11E10CB3786460143C963C8530DF891 " ,
} , {
" txid " : " 0F0685CCEE067638629B1CB27111EB0E15E19B75B1F5D368FC10D216D48FF4A5 " ,
" scriptpubkey " : " 001407DB91DA65EF06B385F4EA20BA05FAF286165C0B " ,
} , {
" txid " : " 822466946527F940A53B823C507A319FDC91CCE55E455D916C9FE13B982058FA " ,
" scriptpubkey " : " 00144A94D23CD5A438531AADD86A0237FE11B9EA4E09 " ,
} , {
" txid " : " 383145E40C8A9F45A0409E080DA5861C9E754B1EC8DD5EFA8A84DEB158E61C88 " ,
" scriptpubkey " : " 0014D0EAC62FDCEE2D1881259BE9CDA4C43DE9050DB8 " ,
} , {
" txid " : " D221BE9B7CDB5FDB58B34D59B30304B7C4C2DF9C3BF73A4AE0E0265642FEC560 " ,
" scriptpubkey " : " 0014D0EAC62FDCEE2D1881259BE9CDA4C43DE9050DB8 " ,
} , {
" txid " : " 420F06E91CEE996D8E75E0565D776A96E8959ECA11E799FFE14522C2D43CCFA5 " ,
" scriptpubkey " : " A9149551336F1E360F5AFB977F24CE72C744A82463D187 " ,
} , {
" txid " : " 9F6127316EBED57E7702A4DF19D6FC0EC23A8FAB9BC0D4AD82C29D3F93C525CD " ,
" scriptpubkey " : " 0014E445493A382C798AF195724DFF67DE4C9250AEC6 " ,
}
]
l2 = node_factory . get_node ( node_id = 3 , dbfile = ' pubkey_regen_commitment_point.sqlite3.xz ' )
results = l2 . db_query ( ' SELECT hex(prev_out_tx) AS txid, hex(scriptpubkey) AS script FROM outputs ' )
scripts = [ { ' txid ' : x [ ' txid ' ] , ' scriptpubkey ' : x [ ' script ' ] } for x in results ]
for exp , actual in zip ( script_map_2 , scripts ) :
assert exp == actual
2020-09-09 09:20:53 +02:00
# Also check that the full_channel_id has been filled in
results = l2 . db_query ( ' SELECT hex(full_channel_id) AS cid, hex(funding_tx_id) as txid, funding_tx_outnum FROM channels ' )
def _chan_id ( txid , outnum ) :
chanid = bytearray . fromhex ( txid )
chanid [ - 1 ] ^ = outnum % 256
chanid [ - 2 ] ^ = outnum / / 256
return chanid . hex ( )
for row in results :
assert _chan_id ( row [ ' txid ' ] , row [ ' funding_tx_outnum ' ] ) == row [ ' cid ' ] . lower ( )
2020-07-29 04:03:03 +02:00
2020-01-07 22:56:08 +01:00
@unittest.skipIf ( VALGRIND and not DEVELOPER , " Without developer valgrind will complain about debug symbols missing " )
2019-12-18 20:00:13 +01:00
def test_optimistic_locking ( node_factory , bitcoind ) :
""" Have a node run against a DB, then change it under its feet, crashing it.
We start a node , wait for it to settle its write so we have a window where
we can interfere , and watch the world burn ( safely ) .
"""
l1 = node_factory . get_node ( may_fail = True , allow_broken_log = True )
sync_blockheight ( bitcoind , [ l1 ] )
l1 . rpc . getinfo ( )
time . sleep ( 1 )
l1 . db . execute ( " UPDATE vars SET intval = intval + 1 WHERE name = ' data_version ' ; " )
# Now trigger any DB write and we should be crashing.
with pytest . raises ( RpcError , match = r ' Connection to RPC server lost. ' ) :
l1 . rpc . newaddr ( )
assert ( l1 . daemon . is_in_log ( r ' Optimistic lock on the database failed ' ) )
2020-09-21 18:34:36 +02:00
@unittest.skipIf ( os . environ . get ( ' TEST_DB_PROVIDER ' , None ) != ' postgres ' , " Only applicable to postgres " )
def test_psql_key_value_dsn ( node_factory , db_provider , monkeypatch ) :
from pyln . testing . db import PostgresDb
# Override get_dsn method to use the key-value style DSN
def get_dsn ( self ) :
print ( " hello " )
return " postgres://host=127.0.0.1 port= {port} user=postgres password=password dbname= {dbname} " . format (
port = self . port , dbname = self . dbname
)
monkeypatch . setattr ( PostgresDb , " get_dsn " , get_dsn )
l1 = node_factory . get_node ( )
opt = [ o for o in l1 . daemon . cmd_line if ' --wallet ' in o ] [ 0 ]
assert ( ' host=127.0.0.1 ' in opt )
2021-02-26 17:12:56 +01:00
@unittest.skipIf (
TEST_NETWORK != ' regtest ' ,
" The DB migration is network specific due to the chain var. "
)
@unittest.skipIf (
os . getenv ( ' TEST_DB_PROVIDER ' , ' sqlite3 ' ) != ' sqlite3 ' ,
" This test is based on a sqlite3 snapshot "
)
def test_local_basepoints_cache ( bitcoind , node_factory ) :
""" XXX started caching the local basepoints as well as the remote ones.
This tests that we can successfully migrate a DB from the
pre - caching state to the caching state , by simply starting the
node up once , issue the HSMd requests , and then store them in the
DB .
"""
# Reestablish the blockheight we had when generating the DB
bitcoind . generate_block ( 6 )
l1 = node_factory . get_node (
dbfile = ' no-local-basepoints.sqlite3.xz ' ,
start = False
)
fields = [
" revocation_basepoint_local " ,
" payment_basepoint_local " ,
" htlc_basepoint_local " ,
" delayed_payment_basepoint_local " ,
]
q = " SELECT {fields} FROM channels " . format ( fields = " , " . join ( fields ) )
# Make sure the DB doesn't have the fields yet.
missing = l1 . db . query ( " SELECT * FROM channels " ) [ 0 ]
for f in fields :
assert ( f not in missing )
# Starting this should cause us to migrate the DB, but none of
# these fields will be set.
l1 . start ( )
present = l1 . db . query ( q ) [ 0 ]
for f in fields :
assert ( f in present )
2021-03-01 15:47:55 +01:00
assert ( present [ f ] is not None )
2021-03-01 19:04:47 +01:00
# New channels should automatically have the basepoints cached.
l2 , l3 = node_factory . line_graph ( 2 )
present = l2 . db . query ( q ) [ 0 ]
for f in fields :
assert ( f in present )
assert ( present [ f ] is not None )
2021-03-01 20:45:39 +01:00
# Restarting will ask hsmd and verify they're unchanged. Remove
# after we verified.
l1 . restart ( )
l2 . restart ( )
2021-10-27 11:49:26 +02:00
@unittest.skipIf ( os . getenv ( ' TEST_DB_PROVIDER ' , ' sqlite3 ' ) != ' sqlite3 ' , " Tests a feature unique to SQLITE3 backend " )
def test_sqlite3_builtin_backup ( bitcoind , node_factory ) :
l1 = node_factory . get_node ( start = False )
# Figure out the path to the actual db.
main_db_file = l1 . db . path
# Create a backup copy in the same location with the suffix .bak
backup_db_file = main_db_file + " .bak "
# Provide the --wallet option and start.
l1 . daemon . opts [ ' wallet ' ] = " sqlite3:// " + main_db_file + ' : ' + backup_db_file
l1 . start ( )
# Get an address and put some funds.
addr = l1 . rpc . newaddr ( ) [ ' bech32 ' ]
bitcoind . rpc . sendtoaddress ( addr , 1 )
bitcoind . generate_block ( 1 )
wait_for ( lambda : len ( l1 . rpc . listfunds ( ) [ ' outputs ' ] ) == 1 )
# Stop the node.
l1 . stop ( )
# Copy the backup over the main db file.
shutil . copyfile ( backup_db_file , main_db_file )
# Remove the --wallet option and start.
del l1 . daemon . opts [ ' wallet ' ]
l1 . start ( )
# Should still see the funds.
assert ( len ( l1 . rpc . listfunds ( ) [ ' outputs ' ] ) == 1 )
2022-07-20 04:28:25 +02:00
@unittest.skipIf ( os . getenv ( ' TEST_DB_PROVIDER ' , ' sqlite3 ' ) != ' sqlite3 ' , " Don ' t know how to swap dbs in Postgres " )
def test_db_sanity_checks ( bitcoind , node_factory ) :
l1 , l2 = node_factory . get_nodes ( 2 , opts = [ { ' allow_broken_log ' : True ,
' may_fail ' : True } , { } ] )
l1 . stop ( )
l2 . stop ( )
# Provide the --wallet option and start with wrong db
l1 . daemon . opts [ ' wallet ' ] = " sqlite3:// " + l2 . db . path
2022-07-21 06:39:30 +02:00
l1 . daemon . start ( wait_for_initialized = False , stderr_redir = True )
2022-07-20 04:28:25 +02:00
l1 . daemon . wait_for_log ( r ' \ * \ *BROKEN \ * \ * wallet: Wallet node_id does not match HSM ' )
# Will have exited with non-zero status.
assert l1 . daemon . proc . wait ( TIMEOUT ) != 0
assert l1 . daemon . is_in_stderr ( ' Wallet sanity check failed ' )
# Now try wrong network,
l1 . daemon . opts [ ' wallet ' ] = " sqlite3:// " + l1 . db . path
l1 . daemon . opts [ ' network ' ] = " bitcoin "
2022-07-21 06:39:30 +02:00
l1 . daemon . start ( wait_for_initialized = False , stderr_redir = True )
2022-07-20 04:28:25 +02:00
l1 . daemon . wait_for_log ( r ' \ * \ *BROKEN \ * \ * wallet: Wallet blockchain hash does not match network blockchain hash ' )
# Will have exited with non-zero status.
assert l1 . daemon . proc . wait ( TIMEOUT ) != 0
assert l1 . daemon . is_in_stderr ( ' Wallet sanity check failed ' )