mirror of
https://gitlab.torproject.org/tpo/core/tor.git
synced 2024-11-20 10:12:15 +01:00
r19049@catbus: nickm | 2008-03-26 12:33:25 -0400
Add new stacklike, free-all-at-once memory allocation strategy. Use it when parsing directory information. This helps parsing speed, and may well help fragmentation some too. hidden-service-related stuff still uses the old tokenizing strategies. svn:r14194
This commit is contained in:
parent
9dfd4132c3
commit
e4ebe3409e
@ -31,6 +31,10 @@ Changes in version 0.2.1.1-alpha - 2008-??-??
|
|||||||
fixed in late 2006. Our new behavior is to call RAND_poll() at
|
fixed in late 2006. Our new behavior is to call RAND_poll() at
|
||||||
startup, and to call RAND_poll() when we reseed later only if we
|
startup, and to call RAND_poll() when we reseed later only if we
|
||||||
have a non-buggy OpenSSL version.
|
have a non-buggy OpenSSL version.
|
||||||
|
- Speed up parsing and cut down on memory fragmentation by using
|
||||||
|
stack-style allocations for parsing directory objects. Previously,
|
||||||
|
this accounted for over 40% of allocations from within Tor's code
|
||||||
|
on a typical directory cache.
|
||||||
- Lots of new unit tests.
|
- Lots of new unit tests.
|
||||||
|
|
||||||
o Code simplifications and refactoring:
|
o Code simplifications and refactoring:
|
||||||
|
@ -9,8 +9,8 @@ else
|
|||||||
libor_extra_source=
|
libor_extra_source=
|
||||||
endif
|
endif
|
||||||
|
|
||||||
libor_a_SOURCES = log.c util.c compat.c container.c mempool.c \
|
libor_a_SOURCES = log.c util.c compat.c container.c mempool.c memarea.c \
|
||||||
$(libor_extra_source)
|
$(libor_extra_source)
|
||||||
libor_crypto_a_SOURCES = crypto.c aes.c tortls.c torgzip.c
|
libor_crypto_a_SOURCES = crypto.c aes.c tortls.c torgzip.c
|
||||||
|
|
||||||
noinst_HEADERS = log.h crypto.h test.h util.h compat.h aes.h torint.h tortls.h strlcpy.c strlcat.c torgzip.h container.h ht.h mempool.h
|
noinst_HEADERS = log.h crypto.h test.h util.h compat.h aes.h torint.h tortls.h strlcpy.c strlcat.c torgzip.h container.h ht.h mempool.h memarea.h
|
||||||
|
211
src/common/memarea.c
Normal file
211
src/common/memarea.c
Normal file
@ -0,0 +1,211 @@
|
|||||||
|
/* Copyright (c) 2008, The Tor Project, Inc. */
|
||||||
|
/* See LICENSE for licensing information */
|
||||||
|
/* $Id$ */
|
||||||
|
|
||||||
|
/** \file memarea.c
|
||||||
|
* \brief Implementation for memarea_t, an allocator for allocating lots of
|
||||||
|
* small objects that will be freed all at once.
|
||||||
|
*/
|
||||||
|
|
||||||
|
#include "orconfig.h"
|
||||||
|
#include <stdlib.h>
|
||||||
|
#include "memarea.h"
|
||||||
|
#include "util.h"
|
||||||
|
#include "compat.h"
|
||||||
|
#include "log.h"
|
||||||
|
|
||||||
|
/** All returned pointers should be aligned to the nearest multiple of this
|
||||||
|
* value. */
|
||||||
|
#define MEMAREA_ALIGN SIZEOF_VOID_P
|
||||||
|
|
||||||
|
#if MEMAREA_ALIGN == 4
|
||||||
|
#define MEMAREA_ALIGN_MASK 3lu
|
||||||
|
#elif MEMAREA_ALIGN == 8
|
||||||
|
#define MEMAREA_ALIGN_MASK 7lu
|
||||||
|
#else
|
||||||
|
#error "void* is neither 4 nor 8 bytes long. I don't know how to align stuff."
|
||||||
|
#endif
|
||||||
|
|
||||||
|
/* Increment <b>ptr</b> until it is aligned to MEMAREA_ALIGN. */
|
||||||
|
static INLINE void *
|
||||||
|
realign_pointer(void *ptr)
|
||||||
|
{
|
||||||
|
uintptr_t x = (uintptr_t)ptr;
|
||||||
|
x = (x+MEMAREA_ALIGN_MASK) & ~MEMAREA_ALIGN_MASK;
|
||||||
|
return (void*)x;
|
||||||
|
}
|
||||||
|
|
||||||
|
/** Implements part of a memarea. New memory is carved off from chunk->mem in
|
||||||
|
* increasing order until a request is too big, at which point a new chunk is
|
||||||
|
* allocated. */
|
||||||
|
typedef struct memarea_chunk_t {
|
||||||
|
/** Next chunk in this area. Only kept around so we can free it. */
|
||||||
|
struct memarea_chunk_t *next_chunk;
|
||||||
|
size_t mem_size; /**< How much RAM is available in u.mem, total? */
|
||||||
|
char *next_mem; /**< Next position in u.mem to allocate data at. If it's
|
||||||
|
* greater than or equal to mem+mem_size, this chunk is
|
||||||
|
* full. */
|
||||||
|
union {
|
||||||
|
char mem[1]; /**< Memory space in this chunk. */
|
||||||
|
void *_void_for_alignment; /**< Dummy; used to make sure mem is aligned. */
|
||||||
|
} u;
|
||||||
|
} memarea_chunk_t;
|
||||||
|
|
||||||
|
#define CHUNK_HEADER_SIZE STRUCT_OFFSET(memarea_chunk_t, u)
|
||||||
|
|
||||||
|
/** A memarea_t is an allocation region for a set of small memory requests
|
||||||
|
* that will all be freed at once. */
|
||||||
|
struct memarea_t {
|
||||||
|
struct memarea_chunk_t *first; /**< Top of the chunk stack: never NULL. */
|
||||||
|
size_t chunk_size; /**<Size to use when allocating chunks.*/
|
||||||
|
};
|
||||||
|
|
||||||
|
/** Helper: allocate a new memarea chunk of around <b>chunk_size</b> bytes. */
|
||||||
|
static memarea_chunk_t *
|
||||||
|
alloc_chunk(size_t chunk_size)
|
||||||
|
{
|
||||||
|
memarea_chunk_t *res = tor_malloc_roundup(&chunk_size);
|
||||||
|
res->next_chunk = NULL;
|
||||||
|
res->mem_size = chunk_size - CHUNK_HEADER_SIZE;
|
||||||
|
res->next_mem = res->u.mem;
|
||||||
|
return res;
|
||||||
|
}
|
||||||
|
|
||||||
|
/** Allocate and return new memarea, with chunks of approximately
|
||||||
|
* <b>chunk_size</b> bytes. (There is indeed some overhead.) */
|
||||||
|
memarea_t *
|
||||||
|
memarea_new(size_t chunk_size)
|
||||||
|
{
|
||||||
|
memarea_t *head = tor_malloc(sizeof(memarea_t));
|
||||||
|
head->first = alloc_chunk(chunk_size);
|
||||||
|
head->chunk_size = chunk_size;
|
||||||
|
return head;
|
||||||
|
}
|
||||||
|
|
||||||
|
/** Free <b>area</b>, invalidating all pointers returned from memarea_alloc()
|
||||||
|
* and friends for this area */
|
||||||
|
void
|
||||||
|
memarea_drop_all(memarea_t *area)
|
||||||
|
{
|
||||||
|
memarea_chunk_t *chunk, *next;
|
||||||
|
for (chunk = area->first; chunk; chunk = next) {
|
||||||
|
next = chunk->next_chunk;
|
||||||
|
tor_free(chunk);
|
||||||
|
}
|
||||||
|
area->first = NULL; /*fail fast on */
|
||||||
|
tor_free(area);
|
||||||
|
}
|
||||||
|
|
||||||
|
/** Forget about having allocated anything in <b>area</b>, and free some of
|
||||||
|
* the backing storage associated with it, as appropriate. Invalidates all
|
||||||
|
* pointers returned from memarea_alloc() for this area. */
|
||||||
|
void
|
||||||
|
memarea_clear(memarea_t *area)
|
||||||
|
{
|
||||||
|
memarea_chunk_t *chunk, *next;
|
||||||
|
if (area->first->next_chunk) {
|
||||||
|
for (chunk = area->first->next_chunk; chunk; chunk = next) {
|
||||||
|
next = chunk->next_chunk;
|
||||||
|
tor_free(chunk);
|
||||||
|
}
|
||||||
|
area->first->next_chunk = NULL;
|
||||||
|
}
|
||||||
|
area->first->next_mem = area->first->u.mem;
|
||||||
|
}
|
||||||
|
|
||||||
|
/** Return true iff <b>p</b> is in a range that has been returned by an
|
||||||
|
* allocation from <b>area</b>. */
|
||||||
|
int
|
||||||
|
memarea_owns_ptr(const memarea_t *area, const void *p)
|
||||||
|
{
|
||||||
|
memarea_chunk_t *chunk;
|
||||||
|
const char *ptr = p;
|
||||||
|
for (chunk = area->first; chunk; chunk = chunk->next_chunk) {
|
||||||
|
if (ptr >= chunk->u.mem && ptr < chunk->next_mem)
|
||||||
|
return 1;
|
||||||
|
}
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
/** Return a pointer to a chunk of memory in <b>area</b> of at least <b>sz</b>
|
||||||
|
* bytes. <b>sz</b> should be significantly smaller than the area's chunk
|
||||||
|
* size, though we can deal if it isn't. */
|
||||||
|
void *
|
||||||
|
memarea_alloc(memarea_t *area, size_t sz)
|
||||||
|
{
|
||||||
|
memarea_chunk_t *chunk = area->first;
|
||||||
|
char *result;
|
||||||
|
tor_assert(chunk);
|
||||||
|
if (chunk->next_mem+sz > chunk->u.mem+chunk->mem_size) {
|
||||||
|
if (sz+CHUNK_HEADER_SIZE >= area->chunk_size) {
|
||||||
|
/* This allocation is too big. Stick it in a special chunk, and put
|
||||||
|
* that chunk second in the list. */
|
||||||
|
memarea_chunk_t *new_chunk = alloc_chunk(sz+CHUNK_HEADER_SIZE);
|
||||||
|
new_chunk->next_chunk = chunk->next_chunk;
|
||||||
|
chunk->next_chunk = new_chunk;
|
||||||
|
chunk = new_chunk;
|
||||||
|
} else {
|
||||||
|
memarea_chunk_t *new_chunk = alloc_chunk(area->chunk_size);
|
||||||
|
new_chunk->next_chunk = chunk;
|
||||||
|
area->first = chunk = new_chunk;
|
||||||
|
}
|
||||||
|
tor_assert(chunk->mem_size >= sz);
|
||||||
|
}
|
||||||
|
result = chunk->next_mem;
|
||||||
|
chunk->next_mem = realign_pointer(chunk->next_mem + sz);
|
||||||
|
return result;
|
||||||
|
}
|
||||||
|
|
||||||
|
/** As memarea_alloc(), but clears the memory it returns. */
|
||||||
|
void *
|
||||||
|
memarea_alloc_zero(memarea_t *area, size_t sz)
|
||||||
|
{
|
||||||
|
void *result = memarea_alloc(area, sz);
|
||||||
|
memset(result, 0, sz);
|
||||||
|
return result;
|
||||||
|
}
|
||||||
|
|
||||||
|
/** As memdup, but returns the memory from <b>area</b>. */
|
||||||
|
void *
|
||||||
|
memarea_memdup(memarea_t *area, const void *s, size_t n)
|
||||||
|
{
|
||||||
|
char *result = memarea_alloc(area, n);
|
||||||
|
memcpy(result, s, n);
|
||||||
|
return result;
|
||||||
|
}
|
||||||
|
|
||||||
|
/** As strdup, but returns the memory from <b>area</b>. */
|
||||||
|
char *
|
||||||
|
memarea_strdup(memarea_t *area, const char *s)
|
||||||
|
{
|
||||||
|
return memarea_memdup(area, s, strlen(s)+1);
|
||||||
|
}
|
||||||
|
|
||||||
|
/** As strndup, but returns the memory from <b>area</b>. */
|
||||||
|
char *
|
||||||
|
memarea_strndup(memarea_t *area, const char *s, size_t n)
|
||||||
|
{
|
||||||
|
size_t ln;
|
||||||
|
char *result;
|
||||||
|
const char *cp, *end = s+n;
|
||||||
|
for (cp = s; *cp && cp < end; ++cp)
|
||||||
|
;
|
||||||
|
/* cp now points to s+n, or to the 0 in the string. */
|
||||||
|
ln = cp-s;
|
||||||
|
result = memarea_memdup(area, s, ln+1);
|
||||||
|
result[ln]='\0';
|
||||||
|
return result;
|
||||||
|
}
|
||||||
|
|
||||||
|
/** Assert that <b>area</b> is okay. */
|
||||||
|
void
|
||||||
|
memarea_assert_ok(memarea_t *area)
|
||||||
|
{
|
||||||
|
memarea_chunk_t *chunk;
|
||||||
|
tor_assert(area->first);
|
||||||
|
|
||||||
|
for (chunk = area->first; chunk; chunk = chunk->next_chunk) {
|
||||||
|
tor_assert(chunk->next_mem >= chunk->u.mem);
|
||||||
|
tor_assert(chunk->next_mem <= chunk->u.mem+chunk->mem_size+MEMAREA_ALIGN);
|
||||||
|
}
|
||||||
|
}
|
22
src/common/memarea.h
Normal file
22
src/common/memarea.h
Normal file
@ -0,0 +1,22 @@
|
|||||||
|
/* Copyright (c) 2008, The Tor Project, Inc. */
|
||||||
|
/* See LICENSE for licensing information */
|
||||||
|
/* $Id$ */
|
||||||
|
/* Tor dependencies */
|
||||||
|
|
||||||
|
#ifndef MEMAREA_H
|
||||||
|
#define MEMAREA_H
|
||||||
|
|
||||||
|
typedef struct memarea_t memarea_t;
|
||||||
|
|
||||||
|
memarea_t *memarea_new(size_t chunk_size);
|
||||||
|
void memarea_drop_all(memarea_t *area);
|
||||||
|
void memarea_clear(memarea_t *area);
|
||||||
|
int memarea_owns_ptr(const memarea_t *area, const void *ptr);
|
||||||
|
void *memarea_alloc(memarea_t *area, size_t sz);
|
||||||
|
void *memarea_alloc_zero(memarea_t *area, size_t sz);
|
||||||
|
void *memarea_memdup(memarea_t *area, const void *s, size_t n);
|
||||||
|
char *memarea_strdup(memarea_t *area, const char *s);
|
||||||
|
char *memarea_strndup(memarea_t *area, const char *s, size_t n);
|
||||||
|
void memarea_assert_ok(memarea_t *area);
|
||||||
|
|
||||||
|
#endif
|
@ -13,6 +13,7 @@ const char routerparse_c_id[] =
|
|||||||
**/
|
**/
|
||||||
|
|
||||||
#include "or.h"
|
#include "or.h"
|
||||||
|
#include "memarea.h"
|
||||||
|
|
||||||
/****************************************************************************/
|
/****************************************************************************/
|
||||||
|
|
||||||
@ -115,7 +116,8 @@ typedef enum {
|
|||||||
*/
|
*/
|
||||||
typedef struct directory_token_t {
|
typedef struct directory_token_t {
|
||||||
directory_keyword tp; /**< Type of the token. */
|
directory_keyword tp; /**< Type of the token. */
|
||||||
int n_args; /**< Number of elements in args */
|
unsigned int in_area:1;/**<XXXX021 remove need for this. */
|
||||||
|
int n_args : 31; /**< Number of elements in args */
|
||||||
char **args; /**< Array of arguments from keyword line. */
|
char **args; /**< Array of arguments from keyword line. */
|
||||||
char *object_type; /**< -----BEGIN [object_type]-----*/
|
char *object_type; /**< -----BEGIN [object_type]-----*/
|
||||||
size_t object_size; /**< Bytes in object_body */
|
size_t object_size; /**< Bytes in object_body */
|
||||||
@ -422,11 +424,13 @@ static directory_token_t *find_first_by_keyword(smartlist_t *s,
|
|||||||
#define TS_ANNOTATIONS_OK 1
|
#define TS_ANNOTATIONS_OK 1
|
||||||
#define TS_NOCHECK 2
|
#define TS_NOCHECK 2
|
||||||
#define TS_NO_NEW_ANNOTATIONS 4
|
#define TS_NO_NEW_ANNOTATIONS 4
|
||||||
static int tokenize_string(const char *start, const char *end,
|
static int tokenize_string(memarea_t *area,
|
||||||
|
const char *start, const char *end,
|
||||||
smartlist_t *out,
|
smartlist_t *out,
|
||||||
token_rule_t *table,
|
token_rule_t *table,
|
||||||
int flags);
|
int flags);
|
||||||
static directory_token_t *get_next_token(const char **s,
|
static directory_token_t *get_next_token(memarea_t *area,
|
||||||
|
const char **s,
|
||||||
const char *eos,
|
const char *eos,
|
||||||
token_rule_t *table);
|
token_rule_t *table);
|
||||||
static int check_signature_token(const char *digest,
|
static int check_signature_token(const char *digest,
|
||||||
@ -629,6 +633,7 @@ router_parse_directory(const char *str)
|
|||||||
const char *end, *cp;
|
const char *end, *cp;
|
||||||
smartlist_t *tokens = NULL;
|
smartlist_t *tokens = NULL;
|
||||||
crypto_pk_env_t *declared_key = NULL;
|
crypto_pk_env_t *declared_key = NULL;
|
||||||
|
memarea_t *area = memarea_new(8192);
|
||||||
|
|
||||||
/* XXXX This could be simplified a lot, but it will all go away
|
/* XXXX This could be simplified a lot, but it will all go away
|
||||||
* once pre-0.1.1.8 is obsolete, and for now it's better not to
|
* once pre-0.1.1.8 is obsolete, and for now it's better not to
|
||||||
@ -649,7 +654,7 @@ router_parse_directory(const char *str)
|
|||||||
}
|
}
|
||||||
++cp;
|
++cp;
|
||||||
tokens = smartlist_create();
|
tokens = smartlist_create();
|
||||||
if (tokenize_string(cp,strchr(cp,'\0'),tokens,dir_token_table,0)) {
|
if (tokenize_string(area,cp,strchr(cp,'\0'),tokens,dir_token_table,0)) {
|
||||||
log_warn(LD_DIR, "Error tokenizing directory signature"); goto err;
|
log_warn(LD_DIR, "Error tokenizing directory signature"); goto err;
|
||||||
}
|
}
|
||||||
if (smartlist_len(tokens) != 1) {
|
if (smartlist_len(tokens) != 1) {
|
||||||
@ -665,8 +670,8 @@ router_parse_directory(const char *str)
|
|||||||
goto err;
|
goto err;
|
||||||
|
|
||||||
SMARTLIST_FOREACH(tokens, directory_token_t *, t, token_free(t));
|
SMARTLIST_FOREACH(tokens, directory_token_t *, t, token_free(t));
|
||||||
smartlist_free(tokens);
|
smartlist_clear(tokens);
|
||||||
tokens = NULL;
|
memarea_clear(area);
|
||||||
|
|
||||||
/* Now try to parse the first part of the directory. */
|
/* Now try to parse the first part of the directory. */
|
||||||
if ((end = strstr(str,"\nrouter "))) {
|
if ((end = strstr(str,"\nrouter "))) {
|
||||||
@ -677,8 +682,7 @@ router_parse_directory(const char *str)
|
|||||||
end = str + strlen(str);
|
end = str + strlen(str);
|
||||||
}
|
}
|
||||||
|
|
||||||
tokens = smartlist_create();
|
if (tokenize_string(area,str,end,tokens,dir_token_table,0)) {
|
||||||
if (tokenize_string(str,end,tokens,dir_token_table,0)) {
|
|
||||||
log_warn(LD_DIR, "Error tokenizing directory"); goto err;
|
log_warn(LD_DIR, "Error tokenizing directory"); goto err;
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -706,6 +710,8 @@ router_parse_directory(const char *str)
|
|||||||
SMARTLIST_FOREACH(tokens, directory_token_t *, t, token_free(t));
|
SMARTLIST_FOREACH(tokens, directory_token_t *, t, token_free(t));
|
||||||
smartlist_free(tokens);
|
smartlist_free(tokens);
|
||||||
}
|
}
|
||||||
|
if (area)
|
||||||
|
memarea_drop_all(area);
|
||||||
return r;
|
return r;
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -722,13 +728,15 @@ router_parse_runningrouters(const char *str)
|
|||||||
crypto_pk_env_t *declared_key = NULL;
|
crypto_pk_env_t *declared_key = NULL;
|
||||||
smartlist_t *tokens = NULL;
|
smartlist_t *tokens = NULL;
|
||||||
const char *eos = str + strlen(str);
|
const char *eos = str + strlen(str);
|
||||||
|
memarea_t *area = NULL;
|
||||||
|
|
||||||
if (router_get_runningrouters_hash(str, digest)) {
|
if (router_get_runningrouters_hash(str, digest)) {
|
||||||
log_warn(LD_DIR, "Unable to compute digest of running-routers");
|
log_warn(LD_DIR, "Unable to compute digest of running-routers");
|
||||||
goto err;
|
goto err;
|
||||||
}
|
}
|
||||||
|
area = memarea_new(8192);
|
||||||
tokens = smartlist_create();
|
tokens = smartlist_create();
|
||||||
if (tokenize_string(str,eos,tokens,dir_token_table,0)) {
|
if (tokenize_string(area,str,eos,tokens,dir_token_table,0)) {
|
||||||
log_warn(LD_DIR, "Error tokenizing running-routers"); goto err;
|
log_warn(LD_DIR, "Error tokenizing running-routers"); goto err;
|
||||||
}
|
}
|
||||||
tok = smartlist_get(tokens,0);
|
tok = smartlist_get(tokens,0);
|
||||||
@ -765,6 +773,8 @@ router_parse_runningrouters(const char *str)
|
|||||||
SMARTLIST_FOREACH(tokens, directory_token_t *, t, token_free(t));
|
SMARTLIST_FOREACH(tokens, directory_token_t *, t, token_free(t));
|
||||||
smartlist_free(tokens);
|
smartlist_free(tokens);
|
||||||
}
|
}
|
||||||
|
if (area)
|
||||||
|
memarea_drop_all(area);
|
||||||
return r;
|
return r;
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -777,6 +787,7 @@ find_dir_signing_key(const char *str, const char *eos)
|
|||||||
const char *cp;
|
const char *cp;
|
||||||
directory_token_t *tok;
|
directory_token_t *tok;
|
||||||
crypto_pk_env_t *key = NULL;
|
crypto_pk_env_t *key = NULL;
|
||||||
|
memarea_t *area = NULL;
|
||||||
tor_assert(str);
|
tor_assert(str);
|
||||||
tor_assert(eos);
|
tor_assert(eos);
|
||||||
|
|
||||||
@ -788,10 +799,11 @@ find_dir_signing_key(const char *str, const char *eos)
|
|||||||
return NULL;
|
return NULL;
|
||||||
++cp; /* Now cp points to the start of the token. */
|
++cp; /* Now cp points to the start of the token. */
|
||||||
|
|
||||||
tok = get_next_token(&cp, eos, dir_token_table);
|
area = memarea_new(1024);
|
||||||
|
tok = get_next_token(area, &cp, eos, dir_token_table);
|
||||||
if (!tok) {
|
if (!tok) {
|
||||||
log_warn(LD_DIR, "Unparseable dir-signing-key token");
|
log_warn(LD_DIR, "Unparseable dir-signing-key token");
|
||||||
return NULL;
|
goto done;
|
||||||
}
|
}
|
||||||
if (tok->tp != K_DIR_SIGNING_KEY) {
|
if (tok->tp != K_DIR_SIGNING_KEY) {
|
||||||
log_warn(LD_DIR, "Dir-signing-key token did not parse as expected");
|
log_warn(LD_DIR, "Dir-signing-key token did not parse as expected");
|
||||||
@ -807,6 +819,7 @@ find_dir_signing_key(const char *str, const char *eos)
|
|||||||
|
|
||||||
done:
|
done:
|
||||||
if (tok) token_free(tok);
|
if (tok) token_free(tok);
|
||||||
|
if (area) memarea_drop_all(area);
|
||||||
return key;
|
return key;
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -1056,6 +1069,7 @@ router_parse_entry_from_string(const char *s, const char *end,
|
|||||||
const char *start_of_annotations, *cp;
|
const char *start_of_annotations, *cp;
|
||||||
size_t prepend_len = prepend_annotations ? strlen(prepend_annotations) : 0;
|
size_t prepend_len = prepend_annotations ? strlen(prepend_annotations) : 0;
|
||||||
int ok = 1;
|
int ok = 1;
|
||||||
|
memarea_t *area = NULL;
|
||||||
|
|
||||||
tor_assert(!allow_annotations || !prepend_annotations);
|
tor_assert(!allow_annotations || !prepend_annotations);
|
||||||
|
|
||||||
@ -1067,9 +1081,10 @@ router_parse_entry_from_string(const char *s, const char *end,
|
|||||||
while (end > s+2 && *(end-1) == '\n' && *(end-2) == '\n')
|
while (end > s+2 && *(end-1) == '\n' && *(end-2) == '\n')
|
||||||
--end;
|
--end;
|
||||||
|
|
||||||
|
area = memarea_new(8192);
|
||||||
tokens = smartlist_create();
|
tokens = smartlist_create();
|
||||||
if (prepend_annotations) {
|
if (prepend_annotations) {
|
||||||
if (tokenize_string(prepend_annotations,NULL,tokens,
|
if (tokenize_string(area,prepend_annotations,NULL,tokens,
|
||||||
routerdesc_token_table,TS_NOCHECK)) {
|
routerdesc_token_table,TS_NOCHECK)) {
|
||||||
log_warn(LD_DIR, "Error tokenizing router descriptor (annotations).");
|
log_warn(LD_DIR, "Error tokenizing router descriptor (annotations).");
|
||||||
goto err;
|
goto err;
|
||||||
@ -1088,7 +1103,7 @@ router_parse_entry_from_string(const char *s, const char *end,
|
|||||||
}
|
}
|
||||||
|
|
||||||
if (allow_annotations && start_of_annotations != s) {
|
if (allow_annotations && start_of_annotations != s) {
|
||||||
if (tokenize_string(start_of_annotations,s,tokens,
|
if (tokenize_string(area,start_of_annotations,s,tokens,
|
||||||
routerdesc_token_table,TS_NOCHECK)) {
|
routerdesc_token_table,TS_NOCHECK)) {
|
||||||
log_warn(LD_DIR, "Error tokenizing router descriptor (annotations).");
|
log_warn(LD_DIR, "Error tokenizing router descriptor (annotations).");
|
||||||
goto err;
|
goto err;
|
||||||
@ -1106,7 +1121,7 @@ router_parse_entry_from_string(const char *s, const char *end,
|
|||||||
if (prepend_annotations)
|
if (prepend_annotations)
|
||||||
flags |= TS_ANNOTATIONS_OK|TS_NO_NEW_ANNOTATIONS;
|
flags |= TS_ANNOTATIONS_OK|TS_NO_NEW_ANNOTATIONS;
|
||||||
|
|
||||||
if (tokenize_string(s,end,tokens,routerdesc_token_table, flags)) {
|
if (tokenize_string(area,s,end,tokens,routerdesc_token_table, flags)) {
|
||||||
log_warn(LD_DIR, "Error tokenizing router descriptor.");
|
log_warn(LD_DIR, "Error tokenizing router descriptor.");
|
||||||
goto err;
|
goto err;
|
||||||
}
|
}
|
||||||
@ -1196,7 +1211,7 @@ router_parse_entry_from_string(const char *s, const char *end,
|
|||||||
}
|
}
|
||||||
|
|
||||||
if ((tok = find_first_by_keyword(tokens, A_PURPOSE))) {
|
if ((tok = find_first_by_keyword(tokens, A_PURPOSE))) {
|
||||||
tor_assert(tok->n_args);
|
tor_assert(tok->n_args != 0);
|
||||||
router->purpose = router_purpose_from_string(tok->args[0]);
|
router->purpose = router_purpose_from_string(tok->args[0]);
|
||||||
} else {
|
} else {
|
||||||
router->purpose = ROUTER_PURPOSE_GENERAL;
|
router->purpose = ROUTER_PURPOSE_GENERAL;
|
||||||
@ -1347,6 +1362,8 @@ router_parse_entry_from_string(const char *s, const char *end,
|
|||||||
if (exit_policy_tokens) {
|
if (exit_policy_tokens) {
|
||||||
smartlist_free(exit_policy_tokens);
|
smartlist_free(exit_policy_tokens);
|
||||||
}
|
}
|
||||||
|
if (area)
|
||||||
|
memarea_drop_all(area);
|
||||||
return router;
|
return router;
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -1366,6 +1383,7 @@ extrainfo_parse_entry_from_string(const char *s, const char *end,
|
|||||||
directory_token_t *tok;
|
directory_token_t *tok;
|
||||||
crypto_pk_env_t *key = NULL;
|
crypto_pk_env_t *key = NULL;
|
||||||
routerinfo_t *router = NULL;
|
routerinfo_t *router = NULL;
|
||||||
|
memarea_t *area = NULL;
|
||||||
|
|
||||||
if (!end) {
|
if (!end) {
|
||||||
end = s + strlen(s);
|
end = s + strlen(s);
|
||||||
@ -1380,7 +1398,8 @@ extrainfo_parse_entry_from_string(const char *s, const char *end,
|
|||||||
return NULL;
|
return NULL;
|
||||||
}
|
}
|
||||||
tokens = smartlist_create();
|
tokens = smartlist_create();
|
||||||
if (tokenize_string(s,end,tokens,extrainfo_token_table,0)) {
|
area = memarea_new(8192);
|
||||||
|
if (tokenize_string(area,s,end,tokens,extrainfo_token_table,0)) {
|
||||||
log_warn(LD_DIR, "Error tokenizing extra-info document.");
|
log_warn(LD_DIR, "Error tokenizing extra-info document.");
|
||||||
goto err;
|
goto err;
|
||||||
}
|
}
|
||||||
@ -1463,6 +1482,8 @@ extrainfo_parse_entry_from_string(const char *s, const char *end,
|
|||||||
SMARTLIST_FOREACH(tokens, directory_token_t *, t, token_free(t));
|
SMARTLIST_FOREACH(tokens, directory_token_t *, t, token_free(t));
|
||||||
smartlist_free(tokens);
|
smartlist_free(tokens);
|
||||||
}
|
}
|
||||||
|
if (area)
|
||||||
|
memarea_drop_all(area);
|
||||||
return extrainfo;
|
return extrainfo;
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -1479,6 +1500,7 @@ authority_cert_parse_from_string(const char *s, const char **end_of_string)
|
|||||||
char *eos;
|
char *eos;
|
||||||
size_t len;
|
size_t len;
|
||||||
int found;
|
int found;
|
||||||
|
memarea_t *area = NULL;
|
||||||
|
|
||||||
s = eat_whitespace(s);
|
s = eat_whitespace(s);
|
||||||
eos = strstr(s, "\n-----END SIGNATURE-----\n");
|
eos = strstr(s, "\n-----END SIGNATURE-----\n");
|
||||||
@ -1492,7 +1514,8 @@ authority_cert_parse_from_string(const char *s, const char **end_of_string)
|
|||||||
len = eos - s;
|
len = eos - s;
|
||||||
|
|
||||||
tokens = smartlist_create();
|
tokens = smartlist_create();
|
||||||
if (tokenize_string(s, eos, tokens, dir_key_certificate_table, 0) < 0) {
|
area = memarea_new(8192);
|
||||||
|
if (tokenize_string(area,s, eos, tokens, dir_key_certificate_table, 0) < 0) {
|
||||||
log_warn(LD_DIR, "Error tokenizing key certificate");
|
log_warn(LD_DIR, "Error tokenizing key certificate");
|
||||||
goto err;
|
goto err;
|
||||||
}
|
}
|
||||||
@ -1542,7 +1565,7 @@ authority_cert_parse_from_string(const char *s, const char **end_of_string)
|
|||||||
|
|
||||||
tok = find_first_by_keyword(tokens, K_DIR_ADDRESS);
|
tok = find_first_by_keyword(tokens, K_DIR_ADDRESS);
|
||||||
if (tok) {
|
if (tok) {
|
||||||
tor_assert(tok->n_args);
|
tor_assert(tok->n_args != 0);
|
||||||
if (parse_addr_port(LOG_WARN, tok->args[0], NULL, &cert->addr,
|
if (parse_addr_port(LOG_WARN, tok->args[0], NULL, &cert->addr,
|
||||||
&cert->dir_port)<0) {
|
&cert->dir_port)<0) {
|
||||||
log_warn(LD_DIR, "Couldn't parse dir-address in certificate");
|
log_warn(LD_DIR, "Couldn't parse dir-address in certificate");
|
||||||
@ -1600,11 +1623,15 @@ authority_cert_parse_from_string(const char *s, const char **end_of_string)
|
|||||||
*end_of_string = eat_whitespace(eos);
|
*end_of_string = eat_whitespace(eos);
|
||||||
}
|
}
|
||||||
SMARTLIST_FOREACH(tokens, directory_token_t *, t, token_free(t));
|
SMARTLIST_FOREACH(tokens, directory_token_t *, t, token_free(t));
|
||||||
|
if (area)
|
||||||
|
memarea_drop_all(area);
|
||||||
smartlist_free(tokens);
|
smartlist_free(tokens);
|
||||||
return cert;
|
return cert;
|
||||||
err:
|
err:
|
||||||
authority_cert_free(cert);
|
authority_cert_free(cert);
|
||||||
SMARTLIST_FOREACH(tokens, directory_token_t *, t, token_free(t));
|
SMARTLIST_FOREACH(tokens, directory_token_t *, t, token_free(t));
|
||||||
|
if (area)
|
||||||
|
memarea_drop_all(area);
|
||||||
smartlist_free(tokens);
|
smartlist_free(tokens);
|
||||||
return NULL;
|
return NULL;
|
||||||
}
|
}
|
||||||
@ -1643,7 +1670,8 @@ find_start_of_next_routerstatus(const char *s)
|
|||||||
* make that consensus.
|
* make that consensus.
|
||||||
**/
|
**/
|
||||||
static routerstatus_t *
|
static routerstatus_t *
|
||||||
routerstatus_parse_entry_from_string(const char **s, smartlist_t *tokens,
|
routerstatus_parse_entry_from_string(memarea_t *area,
|
||||||
|
const char **s, smartlist_t *tokens,
|
||||||
networkstatus_t *vote,
|
networkstatus_t *vote,
|
||||||
vote_routerstatus_t *vote_rs,
|
vote_routerstatus_t *vote_rs,
|
||||||
int consensus_method)
|
int consensus_method)
|
||||||
@ -1658,7 +1686,7 @@ routerstatus_parse_entry_from_string(const char **s, smartlist_t *tokens,
|
|||||||
|
|
||||||
eos = find_start_of_next_routerstatus(*s);
|
eos = find_start_of_next_routerstatus(*s);
|
||||||
|
|
||||||
if (tokenize_string(*s, eos, tokens, rtrstatus_token_table,0)) {
|
if (tokenize_string(area,*s, eos, tokens, rtrstatus_token_table,0)) {
|
||||||
log_warn(LD_DIR, "Error tokenizing router status");
|
log_warn(LD_DIR, "Error tokenizing router status");
|
||||||
goto err;
|
goto err;
|
||||||
}
|
}
|
||||||
@ -1776,8 +1804,7 @@ routerstatus_parse_entry_from_string(const char **s, smartlist_t *tokens,
|
|||||||
tor_version_as_new_as(tok->args[0], "0.2.0.8-alpha");
|
tor_version_as_new_as(tok->args[0], "0.2.0.8-alpha");
|
||||||
}
|
}
|
||||||
if (vote_rs) {
|
if (vote_rs) {
|
||||||
vote_rs->version = tok->args[0];
|
vote_rs->version = tor_strdup(tok->args[0]);
|
||||||
tok->args[0] = NULL; /* suppress free() */
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -1792,6 +1819,8 @@ routerstatus_parse_entry_from_string(const char **s, smartlist_t *tokens,
|
|||||||
done:
|
done:
|
||||||
SMARTLIST_FOREACH(tokens, directory_token_t *, t, token_free(t));
|
SMARTLIST_FOREACH(tokens, directory_token_t *, t, token_free(t));
|
||||||
smartlist_clear(tokens);
|
smartlist_clear(tokens);
|
||||||
|
if (area)
|
||||||
|
memarea_clear(area);
|
||||||
*s = eos;
|
*s = eos;
|
||||||
|
|
||||||
return rs;
|
return rs;
|
||||||
@ -1832,14 +1861,16 @@ networkstatus_v2_parse_from_string(const char *s)
|
|||||||
struct in_addr in;
|
struct in_addr in;
|
||||||
directory_token_t *tok;
|
directory_token_t *tok;
|
||||||
int i;
|
int i;
|
||||||
|
memarea_t *area = NULL;
|
||||||
|
|
||||||
if (router_get_networkstatus_v2_hash(s, ns_digest)) {
|
if (router_get_networkstatus_v2_hash(s, ns_digest)) {
|
||||||
log_warn(LD_DIR, "Unable to compute digest of network-status");
|
log_warn(LD_DIR, "Unable to compute digest of network-status");
|
||||||
goto err;
|
goto err;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
area = memarea_new(8192);
|
||||||
eos = find_start_of_next_routerstatus(s);
|
eos = find_start_of_next_routerstatus(s);
|
||||||
if (tokenize_string(s, eos, tokens, netstatus_token_table,0)) {
|
if (tokenize_string(area,s, eos, tokens, netstatus_token_table,0)) {
|
||||||
log_warn(LD_DIR, "Error tokenizing network-status header.");
|
log_warn(LD_DIR, "Error tokenizing network-status header.");
|
||||||
goto err;
|
goto err;
|
||||||
}
|
}
|
||||||
@ -1857,7 +1888,7 @@ networkstatus_v2_parse_from_string(const char *s)
|
|||||||
tok = find_first_by_keyword(tokens, K_DIR_SOURCE);
|
tok = find_first_by_keyword(tokens, K_DIR_SOURCE);
|
||||||
tor_assert(tok);
|
tor_assert(tok);
|
||||||
tor_assert(tok->n_args >= 3);
|
tor_assert(tok->n_args >= 3);
|
||||||
ns->source_address = tok->args[0]; tok->args[0] = NULL;
|
ns->source_address = tor_strdup(tok->args[0]);
|
||||||
if (tor_inet_aton(tok->args[1], &in) == 0) {
|
if (tor_inet_aton(tok->args[1], &in) == 0) {
|
||||||
log_warn(LD_DIR, "Error parsing network-status source address %s",
|
log_warn(LD_DIR, "Error parsing network-status source address %s",
|
||||||
escaped(tok->args[1]));
|
escaped(tok->args[1]));
|
||||||
@ -1873,7 +1904,7 @@ networkstatus_v2_parse_from_string(const char *s)
|
|||||||
|
|
||||||
tok = find_first_by_keyword(tokens, K_FINGERPRINT);
|
tok = find_first_by_keyword(tokens, K_FINGERPRINT);
|
||||||
tor_assert(tok);
|
tor_assert(tok);
|
||||||
tor_assert(tok->n_args);
|
tor_assert(tok->n_args != 0);
|
||||||
if (base16_decode(ns->identity_digest, DIGEST_LEN, tok->args[0],
|
if (base16_decode(ns->identity_digest, DIGEST_LEN, tok->args[0],
|
||||||
strlen(tok->args[0]))) {
|
strlen(tok->args[0]))) {
|
||||||
log_warn(LD_DIR, "Couldn't decode networkstatus fingerprint %s",
|
log_warn(LD_DIR, "Couldn't decode networkstatus fingerprint %s",
|
||||||
@ -1882,9 +1913,8 @@ networkstatus_v2_parse_from_string(const char *s)
|
|||||||
}
|
}
|
||||||
|
|
||||||
if ((tok = find_first_by_keyword(tokens, K_CONTACT))) {
|
if ((tok = find_first_by_keyword(tokens, K_CONTACT))) {
|
||||||
tor_assert(tok->n_args);
|
tor_assert(tok->n_args != 0);
|
||||||
ns->contact = tok->args[0];
|
ns->contact = tor_strdup(tok->args[0]);
|
||||||
tok->args[0] = NULL;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
tok = find_first_by_keyword(tokens, K_DIR_SIGNING_KEY);
|
tok = find_first_by_keyword(tokens, K_DIR_SIGNING_KEY);
|
||||||
@ -1920,16 +1950,14 @@ networkstatus_v2_parse_from_string(const char *s)
|
|||||||
log_warn(LD_DIR, "Missing client-versions on versioning directory");
|
log_warn(LD_DIR, "Missing client-versions on versioning directory");
|
||||||
goto err;
|
goto err;
|
||||||
}
|
}
|
||||||
ns->client_versions = tok->args[0];
|
ns->client_versions = tor_strdup(tok->args[0]);
|
||||||
tok->args[0] = NULL;
|
|
||||||
|
|
||||||
if (!(tok = find_first_by_keyword(tokens, K_SERVER_VERSIONS)) ||
|
if (!(tok = find_first_by_keyword(tokens, K_SERVER_VERSIONS)) ||
|
||||||
tok->n_args<1) {
|
tok->n_args<1) {
|
||||||
log_warn(LD_DIR, "Missing server-versions on versioning directory");
|
log_warn(LD_DIR, "Missing server-versions on versioning directory");
|
||||||
goto err;
|
goto err;
|
||||||
}
|
}
|
||||||
ns->server_versions = tok->args[0];
|
ns->server_versions = tor_strdup(tok->args[0]);
|
||||||
tok->args[0] = NULL;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
tok = find_first_by_keyword(tokens, K_PUBLISHED);
|
tok = find_first_by_keyword(tokens, K_PUBLISHED);
|
||||||
@ -1943,16 +1971,17 @@ networkstatus_v2_parse_from_string(const char *s)
|
|||||||
s = eos;
|
s = eos;
|
||||||
SMARTLIST_FOREACH(tokens, directory_token_t *, t, token_free(t));
|
SMARTLIST_FOREACH(tokens, directory_token_t *, t, token_free(t));
|
||||||
smartlist_clear(tokens);
|
smartlist_clear(tokens);
|
||||||
|
memarea_clear(area);
|
||||||
while (!strcmpstart(s, "r ")) {
|
while (!strcmpstart(s, "r ")) {
|
||||||
routerstatus_t *rs;
|
routerstatus_t *rs;
|
||||||
if ((rs = routerstatus_parse_entry_from_string(&s, tokens, NULL, NULL, 0)))
|
if ((rs = routerstatus_parse_entry_from_string(NULL, &s, tokens, NULL, NULL, 0)))
|
||||||
smartlist_add(ns->entries, rs);
|
smartlist_add(ns->entries, rs);
|
||||||
}
|
}
|
||||||
smartlist_sort(ns->entries, _compare_routerstatus_entries);
|
smartlist_sort(ns->entries, _compare_routerstatus_entries);
|
||||||
smartlist_uniq(ns->entries, _compare_routerstatus_entries,
|
smartlist_uniq(ns->entries, _compare_routerstatus_entries,
|
||||||
_free_duplicate_routerstatus_entry);
|
_free_duplicate_routerstatus_entry);
|
||||||
|
|
||||||
if (tokenize_string(s, NULL, footer_tokens, dir_footer_token_table,0)) {
|
if (tokenize_string(area,s, NULL, footer_tokens, dir_footer_token_table,0)) {
|
||||||
log_warn(LD_DIR, "Error tokenizing network-status footer.");
|
log_warn(LD_DIR, "Error tokenizing network-status footer.");
|
||||||
goto err;
|
goto err;
|
||||||
}
|
}
|
||||||
@ -1982,7 +2011,8 @@ networkstatus_v2_parse_from_string(const char *s)
|
|||||||
smartlist_free(tokens);
|
smartlist_free(tokens);
|
||||||
SMARTLIST_FOREACH(footer_tokens, directory_token_t *, t, token_free(t));
|
SMARTLIST_FOREACH(footer_tokens, directory_token_t *, t, token_free(t));
|
||||||
smartlist_free(footer_tokens);
|
smartlist_free(footer_tokens);
|
||||||
|
if (area)
|
||||||
|
memarea_drop_all(area);
|
||||||
return ns;
|
return ns;
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -2003,14 +2033,16 @@ networkstatus_parse_vote_from_string(const char *s, const char **eos_out,
|
|||||||
int ok;
|
int ok;
|
||||||
struct in_addr in;
|
struct in_addr in;
|
||||||
int i, inorder, n_signatures = 0;
|
int i, inorder, n_signatures = 0;
|
||||||
|
memarea_t *area = NULL, *rs_area = NULL;
|
||||||
|
|
||||||
if (router_get_networkstatus_v3_hash(s, ns_digest)) {
|
if (router_get_networkstatus_v3_hash(s, ns_digest)) {
|
||||||
log_warn(LD_DIR, "Unable to compute digest of network-status");
|
log_warn(LD_DIR, "Unable to compute digest of network-status");
|
||||||
goto err;
|
goto err;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
area = memarea_new(8192);
|
||||||
end_of_header = find_start_of_next_routerstatus(s);
|
end_of_header = find_start_of_next_routerstatus(s);
|
||||||
if (tokenize_string(s, end_of_header, tokens,
|
if (tokenize_string(area, s, end_of_header, tokens,
|
||||||
is_vote ? networkstatus_token_table :
|
is_vote ? networkstatus_token_table :
|
||||||
networkstatus_consensus_token_table, 0)) {
|
networkstatus_consensus_token_table, 0)) {
|
||||||
log_warn(LD_DIR, "Error tokenizing network-status vote header.");
|
log_warn(LD_DIR, "Error tokenizing network-status vote header.");
|
||||||
@ -2032,7 +2064,7 @@ networkstatus_parse_vote_from_string(const char *s, const char **eos_out,
|
|||||||
|
|
||||||
tok = find_first_by_keyword(tokens, K_VOTE_STATUS);
|
tok = find_first_by_keyword(tokens, K_VOTE_STATUS);
|
||||||
tor_assert(tok);
|
tor_assert(tok);
|
||||||
tor_assert(tok->n_args);
|
tor_assert(tok->n_args != 0);
|
||||||
if (!strcmp(tok->args[0], "vote")) {
|
if (!strcmp(tok->args[0], "vote")) {
|
||||||
ns->is_vote = 1;
|
ns->is_vote = 1;
|
||||||
} else if (!strcmp(tok->args[0], "consensus")) {
|
} else if (!strcmp(tok->args[0], "consensus")) {
|
||||||
@ -2056,8 +2088,7 @@ networkstatus_parse_vote_from_string(const char *s, const char **eos_out,
|
|||||||
tok = find_first_by_keyword(tokens, K_CONSENSUS_METHODS);
|
tok = find_first_by_keyword(tokens, K_CONSENSUS_METHODS);
|
||||||
if (tok) {
|
if (tok) {
|
||||||
for (i=0; i < tok->n_args; ++i)
|
for (i=0; i < tok->n_args; ++i)
|
||||||
smartlist_add(ns->supported_methods, tok->args[i]);
|
smartlist_add(ns->supported_methods, tor_strdup(tok->args[i]));
|
||||||
tok->n_args = 0; /* Prevent double free. */
|
|
||||||
} else {
|
} else {
|
||||||
smartlist_add(ns->supported_methods, tor_strdup("1"));
|
smartlist_add(ns->supported_methods, tor_strdup("1"));
|
||||||
}
|
}
|
||||||
@ -2113,25 +2144,22 @@ networkstatus_parse_vote_from_string(const char *s, const char **eos_out,
|
|||||||
}
|
}
|
||||||
|
|
||||||
if ((tok = find_first_by_keyword(tokens, K_CLIENT_VERSIONS))) {
|
if ((tok = find_first_by_keyword(tokens, K_CLIENT_VERSIONS))) {
|
||||||
ns->client_versions = tok->args[0];
|
ns->client_versions = tor_strdup(tok->args[0]);
|
||||||
tok->args[0] = NULL;
|
|
||||||
}
|
}
|
||||||
if ((tok = find_first_by_keyword(tokens, K_SERVER_VERSIONS))) {
|
if ((tok = find_first_by_keyword(tokens, K_SERVER_VERSIONS))) {
|
||||||
ns->server_versions = tok->args[0];
|
ns->server_versions = tor_strdup(tok->args[0]);
|
||||||
tok->args[0] = NULL;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
tok = find_first_by_keyword(tokens, K_KNOWN_FLAGS);
|
tok = find_first_by_keyword(tokens, K_KNOWN_FLAGS);
|
||||||
ns->known_flags = smartlist_create();
|
ns->known_flags = smartlist_create();
|
||||||
inorder = 1;
|
inorder = 1;
|
||||||
for (i = 0; i < tok->n_args; ++i) {
|
for (i = 0; i < tok->n_args; ++i) {
|
||||||
smartlist_add(ns->known_flags, tok->args[i]);
|
smartlist_add(ns->known_flags, tor_strdup(tok->args[i]));
|
||||||
if (i>0 && strcmp(tok->args[i-1], tok->args[i])>= 0) {
|
if (i>0 && strcmp(tok->args[i-1], tok->args[i])>= 0) {
|
||||||
log_warn(LD_DIR, "%s >= %s", tok->args[i-1], tok->args[i]);
|
log_warn(LD_DIR, "%s >= %s", tok->args[i-1], tok->args[i]);
|
||||||
inorder = 0;
|
inorder = 0;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
tok->n_args = 0; /* suppress free of args members, but not of args itself. */
|
|
||||||
if (!inorder) {
|
if (!inorder) {
|
||||||
log_warn(LD_DIR, "known-flags not in order");
|
log_warn(LD_DIR, "known-flags not in order");
|
||||||
goto err;
|
goto err;
|
||||||
@ -2215,13 +2243,15 @@ networkstatus_parse_vote_from_string(const char *s, const char **eos_out,
|
|||||||
|
|
||||||
/* Parse routerstatus lines. */
|
/* Parse routerstatus lines. */
|
||||||
rs_tokens = smartlist_create();
|
rs_tokens = smartlist_create();
|
||||||
|
rs_area = memarea_new(512);
|
||||||
s = end_of_header;
|
s = end_of_header;
|
||||||
ns->routerstatus_list = smartlist_create();
|
ns->routerstatus_list = smartlist_create();
|
||||||
|
|
||||||
while (!strcmpstart(s, "r ")) {
|
while (!strcmpstart(s, "r ")) {
|
||||||
if (is_vote) {
|
if (is_vote) {
|
||||||
vote_routerstatus_t *rs = tor_malloc_zero(sizeof(vote_routerstatus_t));
|
vote_routerstatus_t *rs = tor_malloc_zero(sizeof(vote_routerstatus_t));
|
||||||
if (routerstatus_parse_entry_from_string(&s, tokens, ns, rs, 0))
|
if (routerstatus_parse_entry_from_string(rs_area, &s, rs_tokens, ns,
|
||||||
|
rs, 0))
|
||||||
smartlist_add(ns->routerstatus_list, rs);
|
smartlist_add(ns->routerstatus_list, rs);
|
||||||
else {
|
else {
|
||||||
tor_free(rs->version);
|
tor_free(rs->version);
|
||||||
@ -2229,7 +2259,8 @@ networkstatus_parse_vote_from_string(const char *s, const char **eos_out,
|
|||||||
}
|
}
|
||||||
} else {
|
} else {
|
||||||
routerstatus_t *rs;
|
routerstatus_t *rs;
|
||||||
if ((rs = routerstatus_parse_entry_from_string(&s, tokens, NULL, NULL,
|
if ((rs = routerstatus_parse_entry_from_string(rs_area, &s, rs_tokens,
|
||||||
|
NULL, NULL,
|
||||||
ns->consensus_method)))
|
ns->consensus_method)))
|
||||||
smartlist_add(ns->routerstatus_list, rs);
|
smartlist_add(ns->routerstatus_list, rs);
|
||||||
}
|
}
|
||||||
@ -2257,7 +2288,7 @@ networkstatus_parse_vote_from_string(const char *s, const char **eos_out,
|
|||||||
++end_of_footer;
|
++end_of_footer;
|
||||||
else
|
else
|
||||||
end_of_footer = s + strlen(s);
|
end_of_footer = s + strlen(s);
|
||||||
if (tokenize_string(s, end_of_footer, footer_tokens,
|
if (tokenize_string(area,s, end_of_footer, footer_tokens,
|
||||||
networkstatus_vote_footer_token_table, 0)) {
|
networkstatus_vote_footer_token_table, 0)) {
|
||||||
log_warn(LD_DIR, "Error tokenizing network-status vote footer.");
|
log_warn(LD_DIR, "Error tokenizing network-status vote footer.");
|
||||||
goto err;
|
goto err;
|
||||||
@ -2355,6 +2386,10 @@ networkstatus_parse_vote_from_string(const char *s, const char **eos_out,
|
|||||||
SMARTLIST_FOREACH(footer_tokens, directory_token_t *, t, token_free(t));
|
SMARTLIST_FOREACH(footer_tokens, directory_token_t *, t, token_free(t));
|
||||||
smartlist_free(footer_tokens);
|
smartlist_free(footer_tokens);
|
||||||
}
|
}
|
||||||
|
if (area)
|
||||||
|
memarea_drop_all(area);
|
||||||
|
if (rs_area)
|
||||||
|
memarea_drop_all(rs_area);
|
||||||
|
|
||||||
return ns;
|
return ns;
|
||||||
}
|
}
|
||||||
@ -2364,8 +2399,9 @@ networkstatus_parse_vote_from_string(const char *s, const char **eos_out,
|
|||||||
ns_detached_signatures_t *
|
ns_detached_signatures_t *
|
||||||
networkstatus_parse_detached_signatures(const char *s, const char *eos)
|
networkstatus_parse_detached_signatures(const char *s, const char *eos)
|
||||||
{
|
{
|
||||||
/* XXXX there is too much duplicate code here. */
|
/* XXXX021 there is too much duplicate code here. */
|
||||||
directory_token_t *tok;
|
directory_token_t *tok;
|
||||||
|
memarea_t *area = NULL;
|
||||||
|
|
||||||
smartlist_t *tokens = smartlist_create();
|
smartlist_t *tokens = smartlist_create();
|
||||||
ns_detached_signatures_t *sigs =
|
ns_detached_signatures_t *sigs =
|
||||||
@ -2374,7 +2410,8 @@ networkstatus_parse_detached_signatures(const char *s, const char *eos)
|
|||||||
if (!eos)
|
if (!eos)
|
||||||
eos = s + strlen(s);
|
eos = s + strlen(s);
|
||||||
|
|
||||||
if (tokenize_string(s, eos, tokens,
|
area = memarea_new(8192);
|
||||||
|
if (tokenize_string(area,s, eos, tokens,
|
||||||
networkstatus_detached_signature_token_table, 0)) {
|
networkstatus_detached_signature_token_table, 0)) {
|
||||||
log_warn(LD_DIR, "Error tokenizing detached networkstatus signatures");
|
log_warn(LD_DIR, "Error tokenizing detached networkstatus signatures");
|
||||||
goto err;
|
goto err;
|
||||||
@ -2463,6 +2500,8 @@ networkstatus_parse_detached_signatures(const char *s, const char *eos)
|
|||||||
done:
|
done:
|
||||||
SMARTLIST_FOREACH(tokens, directory_token_t *, t, token_free(t));
|
SMARTLIST_FOREACH(tokens, directory_token_t *, t, token_free(t));
|
||||||
smartlist_free(tokens);
|
smartlist_free(tokens);
|
||||||
|
if (area)
|
||||||
|
memarea_drop_all(area);
|
||||||
return sigs;
|
return sigs;
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -2480,6 +2519,7 @@ router_parse_addr_policy_item_from_string(const char *s, int assume_action)
|
|||||||
* accept and the address:mask/port element. */
|
* accept and the address:mask/port element. */
|
||||||
char line[TOR_ADDR_BUF_LEN*2 + 32];
|
char line[TOR_ADDR_BUF_LEN*2 + 32];
|
||||||
addr_policy_t *r;
|
addr_policy_t *r;
|
||||||
|
memarea_t *area = NULL;
|
||||||
|
|
||||||
s = eat_whitespace(s);
|
s = eat_whitespace(s);
|
||||||
if ((*s == '*' || TOR_ISDIGIT(*s)) && assume_action >= 0) {
|
if ((*s == '*' || TOR_ISDIGIT(*s)) && assume_action >= 0) {
|
||||||
@ -2495,7 +2535,8 @@ router_parse_addr_policy_item_from_string(const char *s, int assume_action)
|
|||||||
}
|
}
|
||||||
|
|
||||||
eos = cp + strlen(cp);
|
eos = cp + strlen(cp);
|
||||||
tok = get_next_token(&cp, eos, routerdesc_token_table);
|
area = memarea_new(512);
|
||||||
|
tok = get_next_token(area, &cp, eos, routerdesc_token_table);
|
||||||
if (tok->tp == _ERR) {
|
if (tok->tp == _ERR) {
|
||||||
log_warn(LD_DIR, "Error reading address policy: %s", tok->error);
|
log_warn(LD_DIR, "Error reading address policy: %s", tok->error);
|
||||||
goto err;
|
goto err;
|
||||||
@ -2512,6 +2553,8 @@ router_parse_addr_policy_item_from_string(const char *s, int assume_action)
|
|||||||
r = NULL;
|
r = NULL;
|
||||||
done:
|
done:
|
||||||
token_free(tok);
|
token_free(tok);
|
||||||
|
if (area)
|
||||||
|
memarea_drop_all(area);
|
||||||
return r;
|
return r;
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -2620,31 +2663,40 @@ token_free(directory_token_t *tok)
|
|||||||
{
|
{
|
||||||
int i;
|
int i;
|
||||||
tor_assert(tok);
|
tor_assert(tok);
|
||||||
if (tok->args) {
|
|
||||||
for (i = 0; i < tok->n_args; ++i) {
|
|
||||||
tor_free(tok->args[i]);
|
|
||||||
}
|
|
||||||
tor_free(tok->args);
|
|
||||||
}
|
|
||||||
tor_free(tok->object_type);
|
|
||||||
tor_free(tok->object_body);
|
|
||||||
tor_free(tok->error);
|
|
||||||
if (tok->key)
|
if (tok->key)
|
||||||
crypto_free_pk_env(tok->key);
|
crypto_free_pk_env(tok->key);
|
||||||
tor_free(tok);
|
|
||||||
|
if (! tok->in_area) {
|
||||||
|
if (tok->args) {
|
||||||
|
for (i = 0; i < tok->n_args; ++i) {
|
||||||
|
tor_free(tok->args[i]);
|
||||||
|
}
|
||||||
|
tor_free(tok->args);
|
||||||
|
}
|
||||||
|
tor_free(tok->object_type);
|
||||||
|
tor_free(tok->object_body);
|
||||||
|
tor_free(tok->error);
|
||||||
|
tor_free(tok);
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
#define ALLOC_ZERO(sz) (area?memarea_alloc_zero(area,sz):tor_malloc_zero(sz))
|
||||||
|
#define ALLOC(sz) (area?memarea_alloc(area,sz):tor_malloc(sz))
|
||||||
|
#define STRDUP(str) (area?memarea_strdup(area,str):tor_strdup(str))
|
||||||
|
#define STRNDUP(str,n) (area?memarea_strndup(area,(str),(n)):tor_strndup((str),(n)))
|
||||||
|
|
||||||
#define RET_ERR(msg) \
|
#define RET_ERR(msg) \
|
||||||
STMT_BEGIN \
|
STMT_BEGIN \
|
||||||
if (tok) token_free(tok); \
|
if (tok) token_free(tok); \
|
||||||
tok = tor_malloc_zero(sizeof(directory_token_t)); \
|
tok = ALLOC_ZERO(sizeof(directory_token_t)); \
|
||||||
|
tok->in_area = area?1:0; \
|
||||||
tok->tp = _ERR; \
|
tok->tp = _ERR; \
|
||||||
tok->error = tor_strdup(msg); \
|
tok->error = STRDUP(msg); \
|
||||||
goto done_tokenizing; \
|
goto done_tokenizing; \
|
||||||
STMT_END
|
STMT_END
|
||||||
|
|
||||||
static INLINE directory_token_t *
|
static INLINE directory_token_t *
|
||||||
token_check_object(const char *kwd,
|
token_check_object(memarea_t *area, const char *kwd,
|
||||||
directory_token_t *tok, obj_syntax o_syn)
|
directory_token_t *tok, obj_syntax o_syn)
|
||||||
{
|
{
|
||||||
char ebuf[128];
|
char ebuf[128];
|
||||||
@ -2690,7 +2742,8 @@ token_check_object(const char *kwd,
|
|||||||
* of tokens in <b>table</b>.
|
* of tokens in <b>table</b>.
|
||||||
*/
|
*/
|
||||||
static directory_token_t *
|
static directory_token_t *
|
||||||
get_next_token(const char **s, const char *eos, token_rule_t *table)
|
get_next_token(memarea_t *area,
|
||||||
|
const char **s, const char *eos, token_rule_t *table)
|
||||||
{
|
{
|
||||||
const char *next, *eol, *obstart;
|
const char *next, *eol, *obstart;
|
||||||
int i, j, allocated;
|
int i, j, allocated;
|
||||||
@ -2700,16 +2753,8 @@ get_next_token(const char **s, const char *eos, token_rule_t *table)
|
|||||||
char ebuf[128];
|
char ebuf[128];
|
||||||
const char *kwd = "";
|
const char *kwd = "";
|
||||||
|
|
||||||
#define RET_ERR(msg) \
|
tok = ALLOC_ZERO(sizeof(directory_token_t));
|
||||||
STMT_BEGIN \
|
tok->in_area = area?1:0;
|
||||||
if (tok) token_free(tok); \
|
|
||||||
tok = tor_malloc_zero(sizeof(directory_token_t)); \
|
|
||||||
tok->tp = _ERR; \
|
|
||||||
tok->error = tor_strdup(msg); \
|
|
||||||
goto done_tokenizing; \
|
|
||||||
STMT_END
|
|
||||||
|
|
||||||
tok = tor_malloc_zero(sizeof(directory_token_t));
|
|
||||||
tok->tp = _ERR;
|
tok->tp = _ERR;
|
||||||
|
|
||||||
/* Set *s to first token, eol to end-of-line, next to after first token */
|
/* Set *s to first token, eol to end-of-line, next to after first token */
|
||||||
@ -2737,29 +2782,58 @@ get_next_token(const char **s, const char *eos, token_rule_t *table)
|
|||||||
tok->tp = table[i].v;
|
tok->tp = table[i].v;
|
||||||
o_syn = table[i].os;
|
o_syn = table[i].os;
|
||||||
*s = eat_whitespace_eos_no_nl(next, eol);
|
*s = eat_whitespace_eos_no_nl(next, eol);
|
||||||
next = find_whitespace_eos(*s, eol);
|
next = find_whitespace_eos(*s, eol); /* XXXX021 lower this. */
|
||||||
/* We go ahead whether there are arguments or not, so that tok->args is
|
/* We go ahead whether there are arguments or not, so that tok->args is
|
||||||
* always set if we want arguments. */
|
* always set if we want arguments. */
|
||||||
if (table[i].concat_args) {
|
if (table[i].concat_args) {
|
||||||
/* The keyword takes the line as a single argument */
|
/* The keyword takes the line as a single argument */
|
||||||
tok->args = tor_malloc(sizeof(char*));
|
tok->args = ALLOC(sizeof(char*));
|
||||||
tok->args[0] = tor_strndup(*s,eol-*s); /* Grab everything on line */
|
tok->args[0] = STRNDUP(*s,eol-*s); /* Grab everything on line */
|
||||||
tok->n_args = 1;
|
tok->n_args = 1;
|
||||||
} else {
|
} else {
|
||||||
/* This keyword takes multiple arguments. */
|
/* This keyword takes multiple arguments. */
|
||||||
j = 0;
|
if (area) {
|
||||||
allocated = 16;
|
char *mem = memarea_strndup(area, *s, eol-*s);
|
||||||
tok->args = tor_malloc(sizeof(char*)*allocated);
|
char *cp = mem;
|
||||||
while (*s < eol) { /* While not at eol, store the next token */
|
j = 0;
|
||||||
if (j == allocated) {
|
while (*cp) {
|
||||||
allocated *= 2;
|
j++;
|
||||||
tok->args = tor_realloc(tok->args,sizeof(char*)*allocated);
|
cp = (char*)find_whitespace(cp);
|
||||||
|
if (!cp || !*cp)
|
||||||
|
break;
|
||||||
|
cp = (char*)eat_whitespace(cp);
|
||||||
}
|
}
|
||||||
tok->args[j++] = tor_strndup(*s, next-*s);
|
tok->n_args = j;
|
||||||
*s = eat_whitespace_eos_no_nl(next, eol); /* eat intra-line ws */
|
if (tok->n_args) {
|
||||||
next = find_whitespace_eos(*s, eol); /* find end of token at *s */
|
tok->args = memarea_alloc(area, sizeof(char*)*tok->n_args);
|
||||||
|
cp = mem;
|
||||||
|
j = 0;
|
||||||
|
while (*cp) {
|
||||||
|
tok->args[j++] = cp;
|
||||||
|
cp = (char*)find_whitespace(cp);
|
||||||
|
if (!cp || !*cp)
|
||||||
|
break;
|
||||||
|
*cp++ = '\0';
|
||||||
|
cp = (char*)eat_whitespace(cp);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
*s = eol;
|
||||||
|
} else {
|
||||||
|
/*XXXX021 make this needless. */
|
||||||
|
j = 0;
|
||||||
|
allocated = 16;
|
||||||
|
tok->args = tor_malloc(sizeof(char*)*allocated);
|
||||||
|
while (*s < eol) { /* While not at eol, store the next token */
|
||||||
|
if (j == allocated) {
|
||||||
|
allocated *= 2;
|
||||||
|
tok->args = tor_realloc(tok->args,sizeof(char*)*allocated);
|
||||||
|
}
|
||||||
|
tok->args[j++] = tor_strndup(*s, next-*s);
|
||||||
|
*s = eat_whitespace_eos_no_nl(next, eol); /* eat intra-line ws */
|
||||||
|
next = find_whitespace_eos(*s, eol); /* find end of token at *s */
|
||||||
|
}
|
||||||
|
tok->n_args = j;
|
||||||
}
|
}
|
||||||
tok->n_args = j;
|
|
||||||
}
|
}
|
||||||
if (tok->n_args < table[i].min_args) {
|
if (tok->n_args < table[i].min_args) {
|
||||||
tor_snprintf(ebuf, sizeof(ebuf), "Too few arguments to %s", kwd);
|
tor_snprintf(ebuf, sizeof(ebuf), "Too few arguments to %s", kwd);
|
||||||
@ -2778,8 +2852,8 @@ get_next_token(const char **s, const char *eos, token_rule_t *table)
|
|||||||
tok->tp = _A_UNKNOWN;
|
tok->tp = _A_UNKNOWN;
|
||||||
else
|
else
|
||||||
tok->tp = K_OPT;
|
tok->tp = K_OPT;
|
||||||
tok->args = tor_malloc(sizeof(char*));
|
tok->args = ALLOC(sizeof(char*));
|
||||||
tok->args[0] = tor_strndup(*s, eol-*s);
|
tok->args[0] = STRNDUP(*s, eol-*s);
|
||||||
tok->n_args = 1;
|
tok->n_args = 1;
|
||||||
o_syn = OBJ_OK;
|
o_syn = OBJ_OK;
|
||||||
}
|
}
|
||||||
@ -2797,7 +2871,7 @@ get_next_token(const char **s, const char *eos, token_rule_t *table)
|
|||||||
strcmp_len(eol-5, "-----", 5)) { /* nuls or invalid endings */
|
strcmp_len(eol-5, "-----", 5)) { /* nuls or invalid endings */
|
||||||
RET_ERR("Malformed object: bad begin line");
|
RET_ERR("Malformed object: bad begin line");
|
||||||
}
|
}
|
||||||
tok->object_type = tor_strndup(*s+11, eol-*s-16);
|
tok->object_type = STRNDUP(*s+11, eol-*s-16);
|
||||||
obname_len = eol-*s-16; /* store objname length here to avoid a strlen() */
|
obname_len = eol-*s-16; /* store objname length here to avoid a strlen() */
|
||||||
*s = eol+1; /* Set *s to possible start of object data (could be eos) */
|
*s = eol+1; /* Set *s to possible start of object data (could be eos) */
|
||||||
|
|
||||||
@ -2825,7 +2899,7 @@ get_next_token(const char **s, const char *eos, token_rule_t *table)
|
|||||||
RET_ERR("Couldn't parse public key.");
|
RET_ERR("Couldn't parse public key.");
|
||||||
} else { /* If it's something else, try to base64-decode it */
|
} else { /* If it's something else, try to base64-decode it */
|
||||||
int r;
|
int r;
|
||||||
tok->object_body = tor_malloc(next-*s); /* really, this is too much RAM. */
|
tok->object_body = ALLOC(next-*s); /* really, this is too much RAM. */
|
||||||
r = base64_decode(tok->object_body, next-*s, *s, next-*s);
|
r = base64_decode(tok->object_body, next-*s, *s, next-*s);
|
||||||
if (r<0)
|
if (r<0)
|
||||||
RET_ERR("Malformed object: bad base64-encoded data");
|
RET_ERR("Malformed object: bad base64-encoded data");
|
||||||
@ -2834,12 +2908,16 @@ get_next_token(const char **s, const char *eos, token_rule_t *table)
|
|||||||
*s = eol;
|
*s = eol;
|
||||||
|
|
||||||
check_object:
|
check_object:
|
||||||
tok = token_check_object(kwd, tok, o_syn);
|
tok = token_check_object(area, kwd, tok, o_syn);
|
||||||
|
|
||||||
done_tokenizing:
|
done_tokenizing:
|
||||||
return tok;
|
return tok;
|
||||||
|
|
||||||
#undef RET_ERR
|
#undef RET_ERR
|
||||||
|
#undef ALLOC
|
||||||
|
#undef ALLOC_ZERO
|
||||||
|
#undef STRDUP
|
||||||
|
#undef STRNDUP
|
||||||
}
|
}
|
||||||
|
|
||||||
/** Read all tokens from a string between <b>start</b> and <b>end</b>, and add
|
/** Read all tokens from a string between <b>start</b> and <b>end</b>, and add
|
||||||
@ -2848,7 +2926,8 @@ get_next_token(const char **s, const char *eos, token_rule_t *table)
|
|||||||
* entire string.
|
* entire string.
|
||||||
*/
|
*/
|
||||||
static int
|
static int
|
||||||
tokenize_string(const char *start, const char *end, smartlist_t *out,
|
tokenize_string(memarea_t *area,
|
||||||
|
const char *start, const char *end, smartlist_t *out,
|
||||||
token_rule_t *table, int flags)
|
token_rule_t *table, int flags)
|
||||||
{
|
{
|
||||||
const char **s;
|
const char **s;
|
||||||
@ -2864,7 +2943,7 @@ tokenize_string(const char *start, const char *end, smartlist_t *out,
|
|||||||
for (i = 0; i < _NIL; ++i)
|
for (i = 0; i < _NIL; ++i)
|
||||||
counts[i] = 0;
|
counts[i] = 0;
|
||||||
while (*s < end && (!tok || tok->tp != _EOF)) {
|
while (*s < end && (!tok || tok->tp != _EOF)) {
|
||||||
tok = get_next_token(s, end, table);
|
tok = get_next_token(area, s, end, table);
|
||||||
if (tok->tp == _ERR) {
|
if (tok->tp == _ERR) {
|
||||||
log_warn(LD_DIR, "parse error: %s", tok->error);
|
log_warn(LD_DIR, "parse error: %s", tok->error);
|
||||||
token_free(tok);
|
token_free(tok);
|
||||||
@ -3253,7 +3332,7 @@ rend_parse_v2_service_descriptor(rend_service_descriptor_t **parsed_out,
|
|||||||
else
|
else
|
||||||
eos = eos + 1;
|
eos = eos + 1;
|
||||||
/* Tokenize descriptor. */
|
/* Tokenize descriptor. */
|
||||||
if (tokenize_string(desc, eos, tokens, desc_token_table, 0)) {
|
if (tokenize_string(NULL, desc, eos, tokens, desc_token_table, 0)) {
|
||||||
log_warn(LD_REND, "Error tokenizing descriptor.");
|
log_warn(LD_REND, "Error tokenizing descriptor.");
|
||||||
goto err;
|
goto err;
|
||||||
}
|
}
|
||||||
@ -3448,7 +3527,7 @@ rend_decrypt_introduction_points(rend_service_descriptor_t *parsed,
|
|||||||
SMARTLIST_FOREACH(tokens, directory_token_t *, t, token_free(t));
|
SMARTLIST_FOREACH(tokens, directory_token_t *, t, token_free(t));
|
||||||
smartlist_clear(tokens);
|
smartlist_clear(tokens);
|
||||||
/* Tokenize string. */
|
/* Tokenize string. */
|
||||||
if (tokenize_string(*current_ipo, eos, tokens, ipo_token_table, 0)) {
|
if (tokenize_string(NULL, *current_ipo, eos, tokens, ipo_token_table, 0)) {
|
||||||
log_warn(LD_REND, "Error tokenizing introduction point.");
|
log_warn(LD_REND, "Error tokenizing introduction point.");
|
||||||
goto err;
|
goto err;
|
||||||
}
|
}
|
||||||
|
@ -42,6 +42,7 @@ const char tor_svn_revision[] = "";
|
|||||||
#include "test.h"
|
#include "test.h"
|
||||||
#include "torgzip.h"
|
#include "torgzip.h"
|
||||||
#include "mempool.h"
|
#include "mempool.h"
|
||||||
|
#include "memarea.h"
|
||||||
|
|
||||||
int have_failed = 0;
|
int have_failed = 0;
|
||||||
|
|
||||||
@ -3364,6 +3365,97 @@ test_util_mempool(void)
|
|||||||
smartlist_free(allocated);
|
smartlist_free(allocated);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
static void
|
||||||
|
test_util_memarea(void)
|
||||||
|
{
|
||||||
|
memarea_t *area = memarea_new(1024);
|
||||||
|
char *p1, *p2, *p3, *p1_orig;
|
||||||
|
int i;
|
||||||
|
test_assert(area);
|
||||||
|
|
||||||
|
p1_orig = p1 = memarea_alloc(area,64);
|
||||||
|
p2 = memarea_alloc_zero(area,52);
|
||||||
|
p3 = memarea_alloc(area,11);
|
||||||
|
|
||||||
|
test_assert(memarea_owns_ptr(area, p1));
|
||||||
|
test_assert(memarea_owns_ptr(area, p2));
|
||||||
|
test_assert(memarea_owns_ptr(area, p3));
|
||||||
|
/* Make sure we left enough space. */
|
||||||
|
test_assert(p1+64 <= p2);
|
||||||
|
test_assert(p2+52 <= p3);
|
||||||
|
/* Make sure we aligned. */
|
||||||
|
test_eq(((uintptr_t)p1) % sizeof(void*), 0);
|
||||||
|
test_eq(((uintptr_t)p2) % sizeof(void*), 0);
|
||||||
|
test_eq(((uintptr_t)p3) % sizeof(void*), 0);
|
||||||
|
test_assert(!memarea_owns_ptr(area, p3+8192));
|
||||||
|
test_assert(!memarea_owns_ptr(area, p3+30));
|
||||||
|
test_assert(tor_mem_is_zero(p2, 52));
|
||||||
|
/* Make sure we don't overalign. */
|
||||||
|
p1 = memarea_alloc(area, 1);
|
||||||
|
p2 = memarea_alloc(area, 1);
|
||||||
|
test_eq(p1+sizeof(void*), p2);
|
||||||
|
{
|
||||||
|
void *ptr = tor_malloc(64);
|
||||||
|
test_assert(!memarea_owns_ptr(area, ptr));
|
||||||
|
tor_free(ptr);
|
||||||
|
}
|
||||||
|
|
||||||
|
/* memarea_memdup */
|
||||||
|
{
|
||||||
|
char *ptr = tor_malloc(64);
|
||||||
|
crypto_rand(ptr, 64);
|
||||||
|
p1 = memarea_memdup(area, ptr, 64);
|
||||||
|
test_assert(p1 != ptr);
|
||||||
|
test_memeq(p1, ptr, 64);
|
||||||
|
tor_free(ptr);
|
||||||
|
}
|
||||||
|
|
||||||
|
/* memarea_strdup. */
|
||||||
|
p1 = memarea_strdup(area,"");
|
||||||
|
p2 = memarea_strdup(area, "abcd");
|
||||||
|
test_assert(p1);
|
||||||
|
test_assert(p2);
|
||||||
|
test_streq(p1, "");
|
||||||
|
test_streq(p2, "abcd");
|
||||||
|
|
||||||
|
/* memarea_strndup. */
|
||||||
|
{
|
||||||
|
const char *s = "Ad ogni porta batte la morte e grida: il nome!";
|
||||||
|
/* (From Turandot, act 3.) */
|
||||||
|
size_t len = strlen(s);
|
||||||
|
p1 = memarea_strndup(area, s, 1000);
|
||||||
|
p2 = memarea_strndup(area, s, 10);
|
||||||
|
test_streq(p1, s);
|
||||||
|
test_assert(p2 >= p1 + len + 1);
|
||||||
|
test_memeq(s, p2, 10);
|
||||||
|
test_eq(p2[10], '\0');
|
||||||
|
p3 = memarea_strndup(area, s, len);
|
||||||
|
test_streq(p3, s);
|
||||||
|
p3 = memarea_strndup(area, s, len-1);
|
||||||
|
test_memeq(s, p3, len-1);
|
||||||
|
test_eq(p3[len-1], '\0');
|
||||||
|
}
|
||||||
|
|
||||||
|
memarea_clear(area);
|
||||||
|
p1 = memarea_alloc(area, 1);
|
||||||
|
test_eq(p1, p1_orig);
|
||||||
|
memarea_clear(area);
|
||||||
|
|
||||||
|
/* Check for running over an area's size. */
|
||||||
|
for (i = 0; i < 512; ++i) {
|
||||||
|
p1 = memarea_alloc(area, crypto_rand_int(5)+1);
|
||||||
|
test_assert(memarea_owns_ptr(area, p1));
|
||||||
|
}
|
||||||
|
memarea_assert_ok(area);
|
||||||
|
/* Make sure we can allocate a too-big object. */
|
||||||
|
p1 = memarea_alloc_zero(area, 9000);
|
||||||
|
p2 = memarea_alloc_zero(area, 16);
|
||||||
|
test_assert(memarea_owns_ptr(area, p1));
|
||||||
|
test_assert(memarea_owns_ptr(area, p2));
|
||||||
|
|
||||||
|
memarea_drop_all(area);
|
||||||
|
}
|
||||||
|
|
||||||
static void
|
static void
|
||||||
test_util_datadir(void)
|
test_util_datadir(void)
|
||||||
{
|
{
|
||||||
@ -3715,6 +3807,7 @@ static struct {
|
|||||||
SUBENT(util, smartlist),
|
SUBENT(util, smartlist),
|
||||||
SUBENT(util, bitarray),
|
SUBENT(util, bitarray),
|
||||||
SUBENT(util, mempool),
|
SUBENT(util, mempool),
|
||||||
|
SUBENT(util, memarea),
|
||||||
SUBENT(util, strmap),
|
SUBENT(util, strmap),
|
||||||
SUBENT(util, control_formats),
|
SUBENT(util, control_formats),
|
||||||
SUBENT(util, pqueue),
|
SUBENT(util, pqueue),
|
||||||
|
Loading…
Reference in New Issue
Block a user