core-lightning/lightningd/log.c
Rusty Russell 73415d35c9 common: don't send trace messages by default, don't ratelimit at all.
We ratelimited DEBUG messages, but that can be annoying and cause us to miss things.
We demoted the worst offenders in the last release, to TRACE level.

Now, only log trace if it's wanted, and never suppress DEBUG.

Changelog-Changed: Logging: we no longer suppress DEBUG messages from subdaemons.
Signed-off-by: Rusty Russell <rusty@rustcorp.com.au>
Fixes: https://github.com/ElementsProject/lightning/issues/7917
2024-12-16 09:48:51 +10:30

1181 lines
31 KiB
C

#include "config.h"
#include <ccan/err/err.h>
#include <ccan/io/io.h>
#include <ccan/read_write_all/read_write_all.h>
#include <ccan/str/hex/hex.h>
#include <ccan/tal/link/link.h>
#include <ccan/tal/str/str.h>
#include <common/configvar.h>
#include <common/json_command.h>
#include <common/json_param.h>
#include <common/memleak.h>
#include <errno.h>
#include <fcntl.h>
#include <lightningd/log.h>
#include <lightningd/notification.h>
#include <signal.h>
#include <stdio.h>
/* What logging level to use if they didn't specify */
#define DEFAULT_LOGLEVEL LOG_INFORM
/* Once we're up and running, this is set up. */
struct logger *crashlog;
struct print_filter {
/* In list log_book->print_filters / log_file->print_filters */
struct list_node list;
const char *prefix;
enum log_level level;
};
struct log_file {
struct list_head print_filters;
FILE *f;
};
struct log_book {
size_t mem_used;
size_t max_mem;
size_t num_entries;
struct list_head print_filters;
/* Non-null once it's been initialized */
enum log_level *default_print_level;
struct timeabs init_time;
/* Our loggers */
struct list_head loggers;
/* Array of log files: one per ld->logfiles[] */
struct log_file **log_files;
bool print_timestamps;
struct log_entry *log;
/* Prefix this to every entry as you output */
const char *prefix;
/* Although log_book will copy log entries to parent log_book
* (the log_book belongs to lightningd), a pointer to lightningd
* is more directly because the notification needs ld->plugins.
*/
struct lightningd *ld;
/* Cache of all node_ids, to avoid multiple copies. */
struct node_id_map *cache;
};
struct logger {
/* Inside log_book->loggers. */
struct list_node list;
struct log_book *log_book;
const struct node_id *default_node_id;
struct log_prefix *prefix;
/* Print log message at >= this level */
enum log_level print_level;
/* For non-trivial setups, we might need to test filters again
* when actually producing output. */
bool need_refiltering;
};
static struct log_prefix *log_prefix_new(const tal_t *ctx,
const char *prefix TAKES)
{
struct log_prefix *lp = tal(ctx, struct log_prefix);
lp->refcnt = 1;
lp->prefix = tal_strdup(lp, prefix);
return lp;
}
static void log_prefix_drop(struct log_prefix *lp)
{
if (--lp->refcnt == 0)
tal_free(lp);
}
static struct log_prefix *log_prefix_get(struct log_prefix *lp)
{
assert(lp->refcnt);
lp->refcnt++;
return lp;
}
/* Avoids duplicate node_id entries. */
struct node_id_cache {
size_t count;
struct node_id node_id;
};
static const struct node_id *node_cache_id(const struct node_id_cache *nc)
{
return &nc->node_id;
}
static bool node_id_cache_eq(const struct node_id_cache *nc,
const struct node_id *node_id)
{
return node_id_eq(&nc->node_id, node_id);
}
HTABLE_DEFINE_TYPE(struct node_id_cache,
node_cache_id, node_id_hash, node_id_cache_eq,
node_id_map);
static const char *level_prefix(enum log_level level)
{
switch (level) {
case LOG_IO_OUT:
case LOG_IO_IN:
return "IO ";
case LOG_TRACE:
return "TRACE ";
case LOG_DBG:
return "DEBUG ";
case LOG_INFORM:
return "INFO ";
case LOG_UNUSUAL:
return "UNUSUAL";
case LOG_BROKEN:
return "**BROKEN**";
}
abort();
}
/* What do these filters say about level to log this entry at? */
static bool filter_level(const struct list_head *print_filters,
const char *prefix,
const char *node_id_str,
enum log_level *level)
{
struct print_filter *i;
list_for_each(print_filters, i, list) {
if (strstr(prefix, i->prefix) || strstr(node_id_str, i->prefix)) {
*level = i->level;
return true;
}
}
return false;
}
/* What's the lowest filtering which could possibly apply? */
static void lowest_filter(const struct list_head *print_filters,
const char *prefix,
const struct node_id *node_id,
enum log_level *level)
{
struct print_filter *i;
const char *node_id_str;
if (node_id)
node_id_str = fmt_node_id(tmpctx, node_id);
else
node_id_str = NULL;
list_for_each(print_filters, i, list) {
bool match;
if (strstr(prefix, i->prefix))
match = true;
else if (node_id_str) {
match = (strstr(node_id_str, i->prefix) != NULL);
} else {
/* Could this possibly match a node_id? */
match = strstarts(i->prefix, "02") || strstarts(i->prefix, "03");
}
if (match && i->level < *level) {
*level = i->level;
}
}
}
static void log_to_files(const char *log_prefix,
const char *entry_prefix,
enum log_level level,
/* The node_id to log under. */
const struct node_id *node_id,
/* Filters to apply, if non-NULL */
const struct list_head *print_filters,
const struct timeabs *time,
const char *str,
const u8 *io,
size_t io_len,
bool print_timestamps,
const enum log_level *default_print_level,
struct log_file **log_files)
{
char tstamp[sizeof("YYYY-mm-ddTHH:MM:SS.nnnZ ")];
char *entry, *nodestr;
bool filtered;
if (print_timestamps) {
char iso8601_msec_fmt[sizeof("YYYY-mm-ddTHH:MM:SS.%03dZ ")];
strftime(iso8601_msec_fmt, sizeof(iso8601_msec_fmt), "%FT%T.%%03dZ ", gmtime(&time->ts.tv_sec));
snprintf(tstamp, sizeof(tstamp), iso8601_msec_fmt, (int) time->ts.tv_nsec / 1000000);
} else
tstamp[0] = '\0';
if (node_id)
nodestr = fmt_node_id(tmpctx, node_id);
else
nodestr = "";
if (level == LOG_IO_IN || level == LOG_IO_OUT) {
const char *dir = level == LOG_IO_IN ? "[IN]" : "[OUT]";
char *hex = tal_hexstr(NULL, io, io_len);
if (!node_id)
entry = tal_fmt(tmpctx, "%s%s%s: %s%s %s\n",
log_prefix, tstamp, entry_prefix, str, dir, hex);
else
entry = tal_fmt(tmpctx, "%s%s%s-%s: %s%s %s\n",
log_prefix, tstamp,
nodestr,
entry_prefix, str, dir, hex);
tal_free(hex);
} else {
if (!node_id)
entry = tal_fmt(tmpctx, "%s%s%s %s: %s\n",
log_prefix, tstamp, level_prefix(level), entry_prefix, str);
else
entry = tal_fmt(tmpctx, "%s%s%s %s-%s: %s\n",
log_prefix, tstamp, level_prefix(level),
nodestr,
entry_prefix, str);
}
/* In complex configurations, we tell loggers to overshare: then we
* need to filter here to see if we really want it. */
filtered = false;
if (print_filters) {
enum log_level filter;
if (filter_level(print_filters,
entry_prefix, nodestr, &filter)) {
if (level < filter)
return;
/* Even if they specify a default filter level of 'INFO', this overrides */
filtered = true;
}
}
/* Default if nothing set is stdout */
if (!log_files) {
fwrite(entry, strlen(entry), 1, stdout);
fflush(stdout);
}
/* We may have to apply per-file filters. */
for (size_t i = 0; i < tal_count(log_files); i++) {
enum log_level filter;
if (!filter_level(&log_files[i]->print_filters,
entry_prefix, nodestr, &filter)) {
/* If we haven't set default yet, only log UNUSUAL */
if (!default_print_level)
filter = LOG_UNUSUAL;
else {
/* If we've filtered it already, it passes */
if (filtered)
filter = level;
else
filter = *default_print_level;
}
}
if (level < filter)
continue;
fwrite(entry, strlen(entry), 1, log_files[i]->f);
fflush(log_files[i]->f);
}
}
static size_t mem_used(const struct log_entry *e)
{
return sizeof(*e) + strlen(e->log) + 1 + tal_count(e->io);
}
/* Threshold (of 1000) to delete */
static u32 delete_threshold(enum log_level level)
{
switch (level) {
/* Delete 90% of log_io */
case LOG_IO_OUT:
case LOG_IO_IN:
return 900;
/* 50% of LOG_TRACE */
case LOG_TRACE:
return 750;
/* 50% of LOG_DBG */
case LOG_DBG:
return 500;
/* 25% of LOG_INFORM */
case LOG_INFORM:
return 250;
/* 5% of LOG_UNUSUAL / LOG_BROKEN */
case LOG_UNUSUAL:
case LOG_BROKEN:
return 50;
}
abort();
}
/* Delete a log entry: returns how many now deleted */
static size_t delete_entry(struct log_book *log, struct log_entry *i)
{
log->mem_used -= mem_used(i);
log->num_entries--;
if (i->nc && --i->nc->count == 0)
tal_free(i->nc);
free(i->log);
log_prefix_drop(i->prefix);
tal_free(i->io);
return 1 + i->skipped;
}
static size_t prune_log(struct log_book *log)
{
size_t skipped = 0, deleted = 0, count = 0, dst = 0, max, tail;
/* Never delete the last 10% (and definitely not last one!). */
tail = log->num_entries / 10 + 1;
max = log->num_entries - tail;
for (count = 0; count < max; count++) {
struct log_entry *i = &log->log[count];
if (pseudorand(1000) > delete_threshold(i->level)) {
i->skipped += skipped;
skipped = 0;
/* Move down if necesary. */
log->log[dst++] = *i;
continue;
}
skipped += delete_entry(log, i);
deleted++;
}
/* Any skipped at tail go on the next entry */
log->log[count].skipped += skipped;
/* Move down the last 10% */
memmove(log->log + dst, log->log + count, tail * sizeof(*log->log));
return deleted;
}
static void destroy_log_book(struct log_book *log)
{
size_t num = log->num_entries;
for (size_t i = 0; i < num; i++)
delete_entry(log, &log->log[i]);
assert(log->num_entries == 0);
assert(log->mem_used == 0);
}
struct log_book *new_log_book(struct lightningd *ld, size_t max_mem)
{
struct log_book *log_book = tal_linkable(tal(NULL, struct log_book));
/* Give a reasonable size for memory limit! */
assert(max_mem > sizeof(struct logger) * 2);
log_book->mem_used = 0;
log_book->num_entries = 0;
log_book->max_mem = max_mem;
log_book->log_files = NULL;
log_book->default_print_level = NULL;
/* We have to allocate this, since we tal_free it on resetting */
log_book->prefix = tal_strdup(log_book, "");
list_head_init(&log_book->print_filters);
list_head_init(&log_book->loggers);
log_book->init_time = time_now();
log_book->ld = ld;
log_book->cache = tal(log_book, struct node_id_map);
node_id_map_init(log_book->cache);
log_book->log = tal_arr(log_book, struct log_entry, 128);
log_book->print_timestamps = true;
tal_add_destructor(log_book, destroy_log_book);
return log_book;
}
/* What's the minimum level to print this prefix and node_id for this
* log book? Saves us marshalling long print lines in most cases. */
static enum log_level print_level(struct log_book *log_book,
const struct log_prefix *lp,
const struct node_id *node_id,
bool *need_refiltering)
{
enum log_level level = *log_book->default_print_level;
bool have_filters = false;
lowest_filter(&log_book->print_filters, lp->prefix, node_id, &level);
if (!list_empty(&log_book->print_filters))
have_filters = true;
/* We need to look into per-file filters as well: might give a
* lower filter! */
for (size_t i = 0; i < tal_count(log_book->log_files); i++) {
lowest_filter(&log_book->log_files[i]->print_filters,
lp->prefix, node_id, &level);
if (!list_empty(&log_book->log_files[i]->print_filters))
have_filters = true;
}
/* Almost any complex array of filters can mean we want to re-check
* when logging. */
if (need_refiltering)
*need_refiltering = have_filters;
return level;
}
static void destroy_logger(struct logger *log)
{
list_del_from(&log->log_book->loggers, &log->list);
}
/* With different entry points */
struct logger *
new_logger(const tal_t *ctx, struct log_book *log_book,
const struct node_id *default_node_id,
const char *fmt, ...)
{
struct logger *log = tal(ctx, struct logger);
va_list ap;
log->log_book = tal_link(log, log_book);
va_start(ap, fmt);
/* Owned by the log book itself, since it can be referenced
* by log entries, too */
log->prefix = log_prefix_new(log->log_book, take(tal_vfmt(NULL, fmt, ap)));
va_end(ap);
log->default_node_id = tal_dup_or_null(log, struct node_id,
default_node_id);
/* Still initializing? Print UNUSUAL / BROKEN messages only */
if (!log->log_book->default_print_level) {
log->print_level = LOG_UNUSUAL;
log->need_refiltering = false;
} else {
log->print_level = print_level(log->log_book,
log->prefix,
default_node_id,
&log->need_refiltering);
}
list_add(&log->log_book->loggers, &log->list);
tal_add_destructor(log, destroy_logger);
return log;
}
const char *log_prefix(const struct logger *log)
{
return log->prefix->prefix;
}
bool log_has_io_logging(const struct logger *log)
{
return print_level(log->log_book, log->prefix, log->default_node_id, NULL) < LOG_TRACE;
}
bool log_has_trace_logging(const struct logger *log)
{
return print_level(log->log_book, log->prefix, log->default_node_id, NULL) < LOG_DBG;
}
/* This may move entry! */
static void add_entry(struct logger *log, struct log_entry **l)
{
log->log_book->mem_used += mem_used(*l);
log->log_book->num_entries++;
if (log->log_book->mem_used > log->log_book->max_mem) {
size_t old_mem = log->log_book->mem_used, deleted;
deleted = prune_log(log->log_book);
/* Will have moved, but will be last entry. */
*l = &log->log_book->log[log->log_book->num_entries-1];
log_debug(log, "Log pruned %zu entries (mem %zu -> %zu)",
deleted, old_mem, log->log_book->mem_used);
}
}
static void destroy_node_id_cache(struct node_id_cache *nc, struct log_book *log_book)
{
node_id_map_del(log_book->cache, nc);
}
static struct log_entry *new_log_entry(struct logger *log, enum log_level level,
const struct node_id *node_id)
{
struct log_entry *l;
if (log->log_book->num_entries == tal_count(log->log_book->log))
tal_resize(&log->log_book->log, tal_count(log->log_book->log) * 2);
l = &log->log_book->log[log->log_book->num_entries];
l->time = time_now();
l->level = level;
l->skipped = 0;
l->prefix = log_prefix_get(log->prefix);
l->io = NULL;
if (!node_id)
node_id = log->default_node_id;
if (node_id) {
l->nc = node_id_map_get(log->log_book->cache, node_id);
if (!l->nc) {
l->nc = tal(log->log_book->cache, struct node_id_cache);
l->nc->count = 0;
l->nc->node_id = *node_id;
node_id_map_add(log->log_book->cache, l->nc);
tal_add_destructor2(l->nc, destroy_node_id_cache,
log->log_book);
}
l->nc->count++;
} else
l->nc = NULL;
return l;
}
static void maybe_print(struct logger *log, const struct log_entry *l)
{
if (l->level >= log->print_level)
log_to_files(log->log_book->prefix, log->prefix->prefix, l->level,
l->nc ? &l->nc->node_id : NULL,
log->need_refiltering ? &log->log_book->print_filters : NULL,
&l->time, l->log,
l->io, tal_bytelen(l->io),
log->log_book->print_timestamps,
log->log_book->default_print_level,
log->log_book->log_files);
}
static void maybe_notify_log(struct logger *log,
const struct log_entry *l)
{
if (l->level >= log->print_level)
notify_log(log->log_book->ld, l);
}
void logv(struct logger *log, enum log_level level,
const struct node_id *node_id,
bool call_notifier,
const char *fmt, va_list ap)
{
int save_errno = errno;
struct log_entry *l = new_log_entry(log, level, node_id);
/* This is WARN_UNUSED_RESULT, because everyone should somehow deal
* with OOM, even though nobody does. */
if (vasprintf(&l->log, fmt, ap) == -1)
abort();
size_t log_len = strlen(l->log);
/* Sanitize any non-printable characters, and replace with '?' */
for (size_t i=0; i<log_len; i++)
if (l->log[i] < ' ' || l->log[i] >= 0x7f)
l->log[i] = '?';
maybe_print(log, l);
maybe_notify_log(log, l);
add_entry(log, &l);
if (call_notifier)
notify_warning(log->log_book->ld, l);
errno = save_errno;
}
void log_io(struct logger *log, enum log_level dir,
const struct node_id *node_id,
const char *str TAKES,
const void *data TAKES, size_t len)
{
int save_errno = errno;
struct log_entry *l = new_log_entry(log, dir, node_id);
assert(dir == LOG_IO_IN || dir == LOG_IO_OUT);
/* Print first, in case we need to truncate. */
if (l->level >= log->print_level)
log_to_files(log->log_book->prefix, log->prefix->prefix, l->level,
l->nc ? &l->nc->node_id : NULL,
log->need_refiltering ? &log->log_book->print_filters : NULL,
&l->time, str,
data, len,
log->log_book->print_timestamps,
log->log_book->default_print_level,
log->log_book->log_files);
/* Save a tal header, by using raw malloc. */
l->log = strdup(str);
if (taken(str))
tal_free(str);
/* Don't immediately fill buffer with giant IOs */
if (len > log->log_book->max_mem / 64) {
l->skipped++;
len = log->log_book->max_mem / 64;
}
/* FIXME: We could save 4 pointers by using a raw allow, but saving
* the length. */
l->io = tal_dup_arr(log->log_book, u8, data, len, 0);
add_entry(log, &l);
errno = save_errno;
}
void log_(struct logger *log, enum log_level level,
const struct node_id *node_id,
bool call_notifier,
const char *fmt, ...)
{
va_list ap;
va_start(ap, fmt);
logv(log, level, node_id, call_notifier, fmt, ap);
va_end(ap);
}
#define log_each_line(log_book, func, arg) \
log_each_line_((log_book), \
typesafe_cb_preargs(void, void *, (func), (arg), \
unsigned int, \
struct timerel, \
enum log_level, \
const struct node_id *, \
const char *, \
const char *, \
const u8 *), (arg))
static void log_each_line_(const struct log_book *log_book,
void (*func)(unsigned int skipped,
struct timerel time,
enum log_level level,
const struct node_id *node_id,
const char *prefix,
const char *log,
const u8 *io,
void *arg),
void *arg)
{
for (size_t i = 0; i < log_book->num_entries; i++) {
const struct log_entry *l = &log_book->log[i];
func(l->skipped, time_between(l->time, log_book->init_time),
l->level, l->nc ? &l->nc->node_id : NULL,
l->prefix->prefix, l->log, l->io, arg);
}
}
struct log_data {
int fd;
const char *prefix;
};
static void log_one_line(unsigned int skipped,
struct timerel diff,
enum log_level level,
const struct node_id *node_id,
const char *prefix,
const char *log,
const u8 *io,
struct log_data *data)
{
char buf[101];
if (skipped) {
snprintf(buf, sizeof(buf), "%s... %u skipped...", data->prefix, skipped);
write_all(data->fd, buf, strlen(buf));
data->prefix = "\n";
}
snprintf(buf, sizeof(buf), "%s+%lu.%09u %s%s: ",
data->prefix,
(unsigned long)diff.ts.tv_sec,
(unsigned)diff.ts.tv_nsec,
prefix,
level == LOG_IO_IN ? "IO_IN"
: level == LOG_IO_OUT ? "IO_OUT"
: level == LOG_TRACE ? "TRACE"
: level == LOG_DBG ? "DEBUG"
: level == LOG_INFORM ? "INFO"
: level == LOG_UNUSUAL ? "UNUSUAL"
: level == LOG_BROKEN ? "BROKEN"
: "**INVALID**");
write_all(data->fd, buf, strlen(buf));
write_all(data->fd, log, strlen(log));
if (level == LOG_IO_IN || level == LOG_IO_OUT) {
size_t off, used, len = tal_count(io);
/* No allocations, may be in signal handler. */
for (off = 0; off < len; off += used) {
used = len - off;
if (hex_str_size(used) > sizeof(buf))
used = hex_data_size(sizeof(buf));
hex_encode(io + off, used, buf, hex_str_size(used));
write_all(data->fd, buf, strlen(buf));
}
}
data->prefix = "\n";
}
static struct log_file *find_log_file(struct log_book *log_book,
const char *fname)
{
assert(tal_count(log_book->log_files)
== tal_count(log_book->ld->logfiles));
for (size_t i = 0; i < tal_count(log_book->log_files); i++) {
if (streq(log_book->ld->logfiles[i], fname))
return log_book->log_files[i];
}
return NULL;
}
char *opt_log_level(const char *arg, struct log_book *log_book)
{
enum log_level level;
int len;
len = strcspn(arg, ":");
if (!log_level_parse(arg, len, &level))
return tal_fmt(tmpctx, "unknown log level %.*s", len, arg);
if (arg[len]) {
struct print_filter *f = tal(log_book, struct print_filter);
f->prefix = arg + len + 1;
f->level = level;
/* :<filename> */
len = strcspn(f->prefix, ":");
if (f->prefix[len]) {
struct log_file *lf;
lf = find_log_file(log_book, f->prefix + len + 1);
if (!lf)
return tal_fmt(tmpctx,
"unknown log file %s",
f->prefix + len + 1);
f->prefix = tal_strndup(f, f->prefix, len);
list_add_tail(&lf->print_filters, &f->list);
} else {
list_add_tail(&log_book->print_filters, &f->list);
}
} else {
tal_free(log_book->default_print_level);
log_book->default_print_level = tal(log_book, enum log_level);
*log_book->default_print_level = level;
}
return NULL;
}
void json_add_opt_log_levels(struct json_stream *response, struct log_book *log_book)
{
struct print_filter *i;
list_for_each(&log_book->print_filters, i, list) {
json_add_str_fmt(response, "log-level", "%s:%s",
log_level_name(i->level), i->prefix);
}
}
static bool show_log_level(char *buf, size_t len, const struct log_book *log_book)
{
enum log_level l;
if (log_book->default_print_level)
l = *log_book->default_print_level;
else
l = DEFAULT_LOGLEVEL;
strncpy(buf, log_level_name(l), len);
return true;
}
static char *arg_log_prefix(const char *arg, struct log_book *log_book)
{
tal_free(log_book->prefix);
log_book->prefix = tal_strdup(log_book, arg);
return NULL;
}
static bool show_log_prefix(char *buf, size_t len, const struct log_book *log_book)
{
strncpy(buf, log_book->prefix, len);
/* Default is empty, so don't print that! */
return !streq(log_book->prefix, "");
}
static int signalfds[2];
static void handle_sighup(int sig)
{
/* Writes a single 0x00 byte to the signalfds pipe. This may fail if
* we're hammered with SIGHUP. We don't care. */
if (write(signalfds[1], "", 1))
;
}
/* Mutual recursion */
static struct io_plan *setup_read(struct io_conn *conn, struct lightningd *ld);
static struct io_plan *rotate_log(struct io_conn *conn, struct lightningd *ld)
{
log_info(ld->log, "Ending log due to SIGHUP");
for (size_t i = 0; i < tal_count(ld->log->log_book->log_files); i++) {
if (streq(ld->logfiles[i], "-"))
continue;
fclose(ld->log->log_book->log_files[i]->f);
ld->log->log_book->log_files[i]->f = fopen(ld->logfiles[i], "a");
if (!ld->log->log_book->log_files[i]->f)
err(1, "failed to reopen log file %s", ld->logfiles[i]);
}
log_info(ld->log, "Started log due to SIGHUP");
return setup_read(conn, ld);
}
static struct io_plan *setup_read(struct io_conn *conn, struct lightningd *ld)
{
/* We read and discard. */
static char discard;
return io_read(conn, &discard, 1, rotate_log, ld);
}
static void setup_log_rotation(struct lightningd *ld)
{
struct sigaction act;
if (pipe(signalfds) != 0)
errx(1, "Pipe for signalfds");
notleak(io_new_conn(ld, signalfds[0], setup_read, ld));
io_fd_block(signalfds[1], false);
memset(&act, 0, sizeof(act));
act.sa_handler = handle_sighup;
/* We do not need any particular flags; the sigaction
* default behavior (EINTR any system calls, pass only
* the signo to the handler, retain the same signal
* handler throughout) is fine with us.
*/
act.sa_flags = 0;
/* Block all signals while handling SIGHUP.
* Without this, e.g. an inopportune SIGCHLD while we
* are doing a `write` to the SIGHUP signal pipe could
* prevent us from sending the byte and performing the
* log rotation in the main loop.
*
* The SIGHUP handler does very little anyway, and
* the blocked signals will get delivered soon after
* the SIGHUP handler returns.
*/
sigfillset(&act.sa_mask);
if (sigaction(SIGHUP, &act, NULL) != 0)
err(1, "Setting up SIGHUP handler");
}
char *arg_log_to_file(const char *arg, struct lightningd *ld)
{
int size;
struct log_file *logf;
if (!ld->logfiles) {
setup_log_rotation(ld);
ld->logfiles = tal_arr(ld, const char *, 0);
ld->log_book->log_files = tal_arr(ld->log_book, struct log_file *, 0);
}
logf = tal(ld->log_book->log_files, struct log_file);
list_head_init(&logf->print_filters);
if (streq(arg, "-"))
logf->f = stdout;
else {
logf->f = fopen(arg, "a");
if (!logf->f)
return tal_fmt(tmpctx, "Failed to open: %s", strerror(errno));
}
tal_arr_expand(&ld->logfiles, tal_strdup(ld->logfiles, arg));
tal_arr_expand(&ld->log_book->log_files, logf);
/* For convenience make a block of empty lines just like Bitcoin Core */
size = ftell(logf->f);
if (size > 0)
fprintf(logf->f, "\n\n\n\n");
log_debug(ld->log, "Opened log file %s", arg);
return NULL;
}
void opt_register_logging(struct lightningd *ld)
{
opt_register_early_arg("--log-level",
opt_log_level, show_log_level, ld->log_book,
"log level (io, debug, info, unusual, broken) [:prefix]");
clnopt_witharg("--log-timestamps", OPT_EARLY|OPT_SHOWBOOL,
opt_set_bool_arg, opt_show_bool,
&ld->log_book->print_timestamps,
"prefix log messages with timestamp");
clnopt_witharg("--log-prefix", OPT_EARLY|OPT_KEEP_WHITESPACE,
arg_log_prefix, show_log_prefix, ld->log_book, "log prefix");
clnopt_witharg("--log-file=<file>",
OPT_EARLY|OPT_MULTI,
arg_log_to_file, NULL, ld,
"Also log to file (- for stdout)");
}
void logging_options_parsed(struct log_book *log_book)
{
struct logger *log;
/* If they didn't set an explicit level, set to info */
if (!log_book->default_print_level) {
log_book->default_print_level = tal(log_book, enum log_level);
*log_book->default_print_level = DEFAULT_LOGLEVEL;
}
/* Set print_levels for each log, depending on filters. */
list_for_each(&log_book->loggers, log, list) {
log->print_level = print_level(log_book,
log->prefix,
log->default_node_id,
&log->need_refiltering);
}
/* Catch up, since before we were only printing BROKEN msgs */
for (size_t i = 0; i < log_book->num_entries; i++) {
const struct log_entry *l = &log_book->log[i];
if (l->level >= print_level(log_book, l->prefix, l->nc ? &l->nc->node_id : NULL, NULL))
log_to_files(log_book->prefix, l->prefix->prefix, l->level,
l->nc ? &l->nc->node_id : NULL,
&log_book->print_filters,
&l->time, l->log,
l->io, tal_bytelen(l->io),
log_book->print_timestamps,
log_book->default_print_level,
log_book->log_files);
}
}
void log_backtrace_print(const char *fmt, ...)
{
va_list ap;
if (!crashlog)
return;
va_start(ap, fmt);
logv(crashlog, LOG_BROKEN, NULL, false, fmt, ap);
va_end(ap);
}
static void log_dump_to_file(int fd, const struct log_book *log_book)
{
char buf[100];
int len;
struct log_data data;
time_t start;
if (log_book->num_entries == 0) {
write_all(fd, "0 bytes:\n\n", strlen("0 bytes:\n\n"));
return;
}
start = log_book->init_time.ts.tv_sec;
len = snprintf(buf, sizeof(buf), "%zu bytes, %s", log_book->mem_used, ctime(&start));
write_all(fd, buf, len);
/* ctime includes \n... WTF? */
data.prefix = "";
data.fd = fd;
log_each_line(log_book, log_one_line, &data);
write_all(fd, "\n\n", strlen("\n\n"));
}
void log_backtrace_exit(void)
{
int fd;
char timebuf[sizeof("YYYYmmddHHMMSS")];
char logfile[sizeof("/tmp/lightning-crash.log.") + sizeof(timebuf)];
struct timeabs time = time_now();
strftime(timebuf, sizeof(timebuf), "%Y%m%d%H%M%S", gmtime(&time.ts.tv_sec));
if (!crashlog)
return;
/* We expect to be in config dir. */
snprintf(logfile, sizeof(logfile), "crash.log.%s", timebuf);
fd = open(logfile, O_WRONLY|O_CREAT|O_TRUNC, 0600);
if (fd < 0) {
snprintf(logfile, sizeof(logfile),
"/tmp/lightning-crash.log.%s", timebuf);
fd = open(logfile, O_WRONLY|O_CREAT|O_TRUNC, 0600);
}
/* Dump entire log. */
if (fd >= 0) {
log_dump_to_file(fd, crashlog->log_book);
close(fd);
fprintf(stderr, "Log dumped in %s\n", logfile);
}
}
void fatal_vfmt(const char *fmt, va_list ap)
{
va_list ap2;
/* You are not allowed to re-use va_lists, so make a copy. */
va_copy(ap2, ap);
vfprintf(stderr, fmt, ap);
fprintf(stderr, "\n");
if (!crashlog)
exit(1);
logv(crashlog, LOG_BROKEN, NULL, true, fmt, ap2);
abort();
/* va_copy() must be matched with va_end(), even if unreachable. */
va_end(ap2);
}
void fatal(const char *fmt, ...)
{
va_list ap;
va_start(ap, fmt);
fatal_vfmt(fmt, ap);
va_end(ap);
}
struct log_info {
enum log_level level;
struct json_stream *response;
unsigned int num_skipped;
/* If non-null, only show messages about this peer */
const struct node_id *node_id;
};
static void add_skipped(struct log_info *info)
{
if (info->num_skipped) {
json_object_start(info->response, NULL);
json_add_string(info->response, "type", "SKIPPED");
json_add_num(info->response, "num_skipped", info->num_skipped);
json_object_end(info->response);
info->num_skipped = 0;
}
}
static void log_to_json(unsigned int skipped,
struct timerel diff,
enum log_level level,
const struct node_id *node_id,
const char *prefix,
const char *log,
const u8 *io,
struct log_info *info)
{
info->num_skipped += skipped;
if (info->node_id) {
if (!node_id || !node_id_eq(node_id, info->node_id))
return;
}
if (level < info->level) {
info->num_skipped++;
return;
}
add_skipped(info);
json_object_start(info->response, NULL);
json_add_string(info->response, "type",
level == LOG_BROKEN ? "BROKEN"
: level == LOG_UNUSUAL ? "UNUSUAL"
: level == LOG_INFORM ? "INFO"
: level == LOG_DBG ? "DEBUG"
: level == LOG_TRACE ? "TRACE"
: level == LOG_IO_IN ? "IO_IN"
: level == LOG_IO_OUT ? "IO_OUT"
: "UNKNOWN");
json_add_timestr(info->response, "time", diff.ts);
if (node_id)
json_add_node_id(info->response, "node_id", node_id);
json_add_string(info->response, "source", prefix);
json_add_string(info->response, "log", log);
if (io)
json_add_hex_talarr(info->response, "data", io);
json_object_end(info->response);
}
void json_add_log(struct json_stream *response,
const struct log_book *log_book,
const struct node_id *node_id,
enum log_level minlevel)
{
struct log_info info;
info.level = minlevel;
info.response = response;
info.num_skipped = 0;
info.node_id = node_id;
json_array_start(info.response, "log");
log_each_line(log_book, log_to_json, &info);
add_skipped(&info);
json_array_end(info.response);
}
struct command_result *param_loglevel(struct command *cmd,
const char *name,
const char *buffer,
const jsmntok_t *tok,
enum log_level **level)
{
*level = tal(cmd, enum log_level);
if (log_level_parse(buffer + tok->start, tok->end - tok->start, *level))
return NULL;
return command_fail_badparam(cmd, name, buffer, tok,
"should be 'io', 'debug', 'info', or "
"'unusual'");
}
static struct command_result *json_getlog(struct command *cmd,
const char *buffer,
const jsmntok_t *obj UNNEEDED,
const jsmntok_t * params)
{
struct json_stream *response;
enum log_level *minlevel;
struct log_book *log_book = cmd->ld->log_book;
if (!param(cmd, buffer, params,
p_opt_def("level", param_loglevel, &minlevel, LOG_INFORM),
NULL))
return command_param_failed();
response = json_stream_success(cmd);
/* Suppress logging for this stream, to not bloat io logs */
json_stream_log_suppress_for_cmd(response, cmd);
json_add_timestr(response, "created_at", log_book->init_time.ts);
json_add_num(response, "bytes_used", (unsigned int)log_book->mem_used);
json_add_num(response, "bytes_max", (unsigned int)log_book->max_mem);
json_add_log(response, log_book, NULL, *minlevel);
return command_success(cmd, response);
}
static const struct json_command getlog_command = {
"getlog",
json_getlog,
};
AUTODATA(json_command, &getlog_command);