Implemented link padding and receiver token buckets
Each socket reads at most 'bandwidth' bytes per second sustained, but
can handle bursts of up to 10*bandwidth bytes.
Cells are now sent out at evenly-spaced intervals, with padding sent
out otherwise. Set Linkpadding=0 in the rc file to send cells as soon
as they're available (and to never send padding cells).
Added license/copyrights statements at the top of most files.
router->min and router->max have been merged into a single 'bandwidth'
value. We should make the routerinfo_t reflect this (want to do that,
Mat?)
As the bandwidth increases, and we want to stop sleeping more and more
frequently to send a single cell, cpu usage goes up. At 128kB/s we're
pretty much calling poll with a timeout of 1ms or even 0ms. The current
code takes a timeout of 0-9ms and makes it 10ms. prepare_for_poll()
handles everything that should have happened in the past, so as long as
our buffers don't get too full in that 10ms, we're ok.
Speaking of too full, if you run three servers at 100kB/s with -l debug,
it spends too much time printing debugging messages to be able to keep
up with the cells. The outbuf ultimately fills up and it kills that
connection. If you run with -l err, it works fine up through 500kB/s and
probably beyond. Down the road we'll want to teach it to recognize when
an outbuf is getting full, and back off.
svn:r50
2002-07-16 03:12:15 +02:00
|
|
|
/* Copyright 2001,2002 Roger Dingledine, Matej Pfajfar. */
|
|
|
|
/* See LICENSE for licensing information */
|
|
|
|
/* $Id$ */
|
2002-06-27 00:45:49 +02:00
|
|
|
|
|
|
|
#include "or.h"
|
|
|
|
|
2003-07-05 07:46:06 +02:00
|
|
|
/********* START PROTOTYPES **********/
|
|
|
|
|
|
|
|
static void dumpstats(void); /* dump stats to stdout */
|
2003-09-25 07:17:11 +02:00
|
|
|
static int init_descriptor(void);
|
2003-07-05 07:46:06 +02:00
|
|
|
|
2002-06-27 00:45:49 +02:00
|
|
|
/********* START VARIABLES **********/
|
|
|
|
|
2003-08-11 22:22:48 +02:00
|
|
|
extern char *conn_type_to_string[];
|
2003-09-25 12:42:07 +02:00
|
|
|
extern char *conn_state_to_string[][_CONN_TYPE_MAX+1];
|
2003-08-11 22:22:48 +02:00
|
|
|
|
Implemented link padding and receiver token buckets
Each socket reads at most 'bandwidth' bytes per second sustained, but
can handle bursts of up to 10*bandwidth bytes.
Cells are now sent out at evenly-spaced intervals, with padding sent
out otherwise. Set Linkpadding=0 in the rc file to send cells as soon
as they're available (and to never send padding cells).
Added license/copyrights statements at the top of most files.
router->min and router->max have been merged into a single 'bandwidth'
value. We should make the routerinfo_t reflect this (want to do that,
Mat?)
As the bandwidth increases, and we want to stop sleeping more and more
frequently to send a single cell, cpu usage goes up. At 128kB/s we're
pretty much calling poll with a timeout of 1ms or even 0ms. The current
code takes a timeout of 0-9ms and makes it 10ms. prepare_for_poll()
handles everything that should have happened in the past, so as long as
our buffers don't get too full in that 10ms, we're ok.
Speaking of too full, if you run three servers at 100kB/s with -l debug,
it spends too much time printing debugging messages to be able to keep
up with the cells. The outbuf ultimately fills up and it kills that
connection. If you run with -l err, it works fine up through 500kB/s and
probably beyond. Down the road we'll want to teach it to recognize when
an outbuf is getting full, and back off.
svn:r50
2002-07-16 03:12:15 +02:00
|
|
|
or_options_t options; /* command-line and config-file options */
|
2003-07-05 09:10:34 +02:00
|
|
|
int global_read_bucket; /* max number of bytes I can read this second */
|
2002-06-27 00:45:49 +02:00
|
|
|
|
2002-07-05 08:27:23 +02:00
|
|
|
static connection_t *connection_array[MAXCONNECTIONS] =
|
2002-06-27 00:45:49 +02:00
|
|
|
{ NULL };
|
|
|
|
|
2002-09-03 20:44:24 +02:00
|
|
|
static struct pollfd poll_array[MAXCONNECTIONS];
|
2002-06-27 00:45:49 +02:00
|
|
|
|
2002-07-05 08:27:23 +02:00
|
|
|
static int nfds=0; /* number of connections currently active */
|
2002-06-27 00:45:49 +02:00
|
|
|
|
2003-08-12 10:04:31 +02:00
|
|
|
#ifndef MS_WINDOWS /* do signal stuff only on unix */
|
2002-09-22 00:41:48 +02:00
|
|
|
static int please_dumpstats=0; /* whether we should dump stats during the loop */
|
2003-09-21 08:15:43 +02:00
|
|
|
static int please_reset =0; /* whether we just got a sighup */
|
2003-08-12 08:41:53 +02:00
|
|
|
static int please_reap_children=0; /* whether we should waitpid for exited children*/
|
2003-08-12 10:04:31 +02:00
|
|
|
#endif /* signal stuff */
|
2002-09-22 00:41:48 +02:00
|
|
|
|
2003-09-25 07:17:11 +02:00
|
|
|
/* private keys */
|
|
|
|
static crypto_pk_env_t *onionkey=NULL;
|
|
|
|
static crypto_pk_env_t *linkkey=NULL;
|
|
|
|
static crypto_pk_env_t *identitykey=NULL;
|
2002-06-27 00:45:49 +02:00
|
|
|
|
2002-09-26 15:17:14 +02:00
|
|
|
routerinfo_t *my_routerinfo=NULL;
|
|
|
|
|
2002-06-27 00:45:49 +02:00
|
|
|
/********* END VARIABLES ************/
|
|
|
|
|
2003-09-25 07:17:11 +02:00
|
|
|
void set_onion_key(crypto_pk_env_t *k) {
|
|
|
|
onionkey = k;
|
2002-09-28 02:52:59 +02:00
|
|
|
}
|
|
|
|
|
2003-09-25 07:17:11 +02:00
|
|
|
crypto_pk_env_t *get_onion_key(void) {
|
|
|
|
assert(onionkey);
|
|
|
|
return onionkey;
|
2002-09-28 02:52:59 +02:00
|
|
|
}
|
|
|
|
|
2003-09-25 07:17:11 +02:00
|
|
|
void set_link_key(crypto_pk_env_t *k)
|
|
|
|
{
|
|
|
|
linkkey = k;
|
|
|
|
}
|
|
|
|
|
|
|
|
crypto_pk_env_t *get_link_key(void)
|
|
|
|
{
|
|
|
|
assert(linkkey);
|
|
|
|
return linkkey;
|
2003-05-08 00:40:03 +02:00
|
|
|
}
|
|
|
|
|
2003-09-25 07:17:11 +02:00
|
|
|
void set_identity_key(crypto_pk_env_t *k) {
|
|
|
|
identitykey = k;
|
|
|
|
}
|
|
|
|
|
|
|
|
crypto_pk_env_t *get_identity_key(void) {
|
|
|
|
assert(identitykey);
|
|
|
|
return identitykey;
|
2003-05-08 00:40:03 +02:00
|
|
|
}
|
|
|
|
|
2002-07-05 08:27:23 +02:00
|
|
|
/****************************************************************************
|
|
|
|
*
|
|
|
|
* This section contains accessors and other methods on the connection_array
|
|
|
|
* and poll_array variables (which are global within this file and unavailable
|
|
|
|
* outside it).
|
|
|
|
*
|
|
|
|
****************************************************************************/
|
|
|
|
|
2002-06-27 00:45:49 +02:00
|
|
|
int connection_add(connection_t *conn) {
|
|
|
|
|
2002-09-03 20:36:40 +02:00
|
|
|
if(nfds >= options.MaxConn-1) {
|
2003-09-26 12:03:50 +02:00
|
|
|
log(LOG_WARNING,"connection_add(): failing because nfds is too high.");
|
2002-06-27 00:45:49 +02:00
|
|
|
return -1;
|
|
|
|
}
|
2003-09-25 07:17:11 +02:00
|
|
|
|
2002-06-27 00:45:49 +02:00
|
|
|
conn->poll_index = nfds;
|
|
|
|
connection_set_poll_socket(conn);
|
|
|
|
connection_array[nfds] = conn;
|
|
|
|
|
|
|
|
/* zero these out here, because otherwise we'll inherit values from the previously freed one */
|
|
|
|
poll_array[nfds].events = 0;
|
|
|
|
poll_array[nfds].revents = 0;
|
|
|
|
|
|
|
|
nfds++;
|
|
|
|
|
2002-07-16 04:12:58 +02:00
|
|
|
log(LOG_INFO,"connection_add(): new conn type %d, socket %d, nfds %d.",conn->type, conn->s, nfds);
|
2002-06-27 00:45:49 +02:00
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
void connection_set_poll_socket(connection_t *conn) {
|
|
|
|
poll_array[conn->poll_index].fd = conn->s;
|
|
|
|
}
|
|
|
|
|
|
|
|
int connection_remove(connection_t *conn) {
|
|
|
|
int current_index;
|
|
|
|
|
|
|
|
assert(conn);
|
|
|
|
assert(nfds>0);
|
|
|
|
|
2002-07-22 06:08:37 +02:00
|
|
|
log(LOG_INFO,"connection_remove(): removing socket %d, nfds now %d",conn->s, nfds-1);
|
major overhaul: dns slave subsystem, topics
on startup, it forks off a master dns handler, which forks off dns
slaves (like the apache model). slaves as spawned as load increases,
and then reused. excess slaves are not ever killed, currently.
implemented topics. each topic has a receive window in each direction
at each edge of the circuit, and sends sendme's at the data level, as
per before. each circuit also has receive windows in each direction at
each hop; an edge sends a circuit-level sendme as soon as enough data
cells have arrived (regardless of whether the data cells were flushed
to the exit conns). removed the 'connected' cell type, since it's now
a topic command within data cells.
at the edge of the circuit, there can be multiple connections associated
with a single circuit. you find them via the linked list conn->next_topic.
currently each new ap connection starts its own circuit, so we ought
to see comparable performance to what we had before. but that's only
because i haven't written the code to reattach to old circuits. please
try to break it as-is, and then i'll make it reuse the same circuit and
we'll try to break that.
svn:r152
2003-01-26 10:02:24 +01:00
|
|
|
circuit_about_to_close_connection(conn); /* if it's an edge conn, remove it from the list
|
|
|
|
* of conn's on this circuit. If it's not on an edge,
|
|
|
|
* flush and send destroys for all circuits on this conn
|
|
|
|
*/
|
2002-06-27 00:45:49 +02:00
|
|
|
|
|
|
|
current_index = conn->poll_index;
|
|
|
|
if(current_index == nfds-1) { /* this is the end */
|
|
|
|
nfds--;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* we replace this one with the one at the end, then free it */
|
|
|
|
nfds--;
|
|
|
|
poll_array[current_index].fd = poll_array[nfds].fd;
|
|
|
|
poll_array[current_index].events = poll_array[nfds].events;
|
|
|
|
poll_array[current_index].revents = poll_array[nfds].revents;
|
|
|
|
connection_array[current_index] = connection_array[nfds];
|
|
|
|
connection_array[current_index]->poll_index = current_index;
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2002-07-08 10:59:15 +02:00
|
|
|
connection_t *connection_twin_get_by_addr_port(uint32_t addr, uint16_t port) {
|
Implemented link padding and receiver token buckets
Each socket reads at most 'bandwidth' bytes per second sustained, but
can handle bursts of up to 10*bandwidth bytes.
Cells are now sent out at evenly-spaced intervals, with padding sent
out otherwise. Set Linkpadding=0 in the rc file to send cells as soon
as they're available (and to never send padding cells).
Added license/copyrights statements at the top of most files.
router->min and router->max have been merged into a single 'bandwidth'
value. We should make the routerinfo_t reflect this (want to do that,
Mat?)
As the bandwidth increases, and we want to stop sleeping more and more
frequently to send a single cell, cpu usage goes up. At 128kB/s we're
pretty much calling poll with a timeout of 1ms or even 0ms. The current
code takes a timeout of 0-9ms and makes it 10ms. prepare_for_poll()
handles everything that should have happened in the past, so as long as
our buffers don't get too full in that 10ms, we're ok.
Speaking of too full, if you run three servers at 100kB/s with -l debug,
it spends too much time printing debugging messages to be able to keep
up with the cells. The outbuf ultimately fills up and it kills that
connection. If you run with -l err, it works fine up through 500kB/s and
probably beyond. Down the road we'll want to teach it to recognize when
an outbuf is getting full, and back off.
svn:r50
2002-07-16 03:12:15 +02:00
|
|
|
/* Find a connection to the router described by addr and port,
|
|
|
|
* or alternately any router which knows its key.
|
|
|
|
* This connection *must* be in 'open' state.
|
|
|
|
* If not, return NULL.
|
|
|
|
*/
|
2002-07-08 10:59:15 +02:00
|
|
|
int i;
|
|
|
|
connection_t *conn;
|
2002-07-19 01:44:57 +02:00
|
|
|
routerinfo_t *router;
|
2002-07-08 10:59:15 +02:00
|
|
|
|
|
|
|
/* first check if it's there exactly */
|
|
|
|
conn = connection_exact_get_by_addr_port(addr,port);
|
2003-06-25 01:14:39 +02:00
|
|
|
if(conn && connection_state_is_open(conn)) {
|
2002-08-23 05:35:44 +02:00
|
|
|
log(LOG_INFO,"connection_twin_get_by_addr_port(): Found exact match.");
|
2002-07-08 10:59:15 +02:00
|
|
|
return conn;
|
Implemented link padding and receiver token buckets
Each socket reads at most 'bandwidth' bytes per second sustained, but
can handle bursts of up to 10*bandwidth bytes.
Cells are now sent out at evenly-spaced intervals, with padding sent
out otherwise. Set Linkpadding=0 in the rc file to send cells as soon
as they're available (and to never send padding cells).
Added license/copyrights statements at the top of most files.
router->min and router->max have been merged into a single 'bandwidth'
value. We should make the routerinfo_t reflect this (want to do that,
Mat?)
As the bandwidth increases, and we want to stop sleeping more and more
frequently to send a single cell, cpu usage goes up. At 128kB/s we're
pretty much calling poll with a timeout of 1ms or even 0ms. The current
code takes a timeout of 0-9ms and makes it 10ms. prepare_for_poll()
handles everything that should have happened in the past, so as long as
our buffers don't get too full in that 10ms, we're ok.
Speaking of too full, if you run three servers at 100kB/s with -l debug,
it spends too much time printing debugging messages to be able to keep
up with the cells. The outbuf ultimately fills up and it kills that
connection. If you run with -l err, it works fine up through 500kB/s and
probably beyond. Down the road we'll want to teach it to recognize when
an outbuf is getting full, and back off.
svn:r50
2002-07-16 03:12:15 +02:00
|
|
|
}
|
2002-07-08 10:59:15 +02:00
|
|
|
|
|
|
|
/* now check if any of the other open connections are a twin for this one */
|
|
|
|
|
2002-07-19 01:44:57 +02:00
|
|
|
router = router_get_by_addr_port(addr,port);
|
|
|
|
if(!router)
|
|
|
|
return NULL;
|
|
|
|
|
|
|
|
for(i=0;i<nfds;i++) {
|
|
|
|
conn = connection_array[i];
|
|
|
|
assert(conn);
|
2003-06-25 01:14:39 +02:00
|
|
|
if(connection_state_is_open(conn) &&
|
|
|
|
!conn->marked_for_close &&
|
2003-09-25 07:17:11 +02:00
|
|
|
!crypto_pk_cmp_keys(conn->onion_pkey, router->onion_pkey)) {
|
2002-08-23 05:35:44 +02:00
|
|
|
log(LOG_INFO,"connection_twin_get_by_addr_port(): Found twin (%s).",conn->address);
|
2002-07-19 01:44:57 +02:00
|
|
|
return conn;
|
|
|
|
}
|
|
|
|
}
|
2002-07-08 10:59:15 +02:00
|
|
|
/* guess not */
|
|
|
|
return NULL;
|
|
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
connection_t *connection_exact_get_by_addr_port(uint32_t addr, uint16_t port) {
|
2002-06-27 00:45:49 +02:00
|
|
|
int i;
|
|
|
|
connection_t *conn;
|
|
|
|
|
|
|
|
for(i=0;i<nfds;i++) {
|
|
|
|
conn = connection_array[i];
|
2003-06-25 01:09:21 +02:00
|
|
|
if(conn->addr == addr && conn->port == port && !conn->marked_for_close)
|
2003-06-25 09:19:30 +02:00
|
|
|
return conn;
|
2002-06-27 00:45:49 +02:00
|
|
|
}
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
|
|
|
connection_t *connection_get_by_type(int type) {
|
|
|
|
int i;
|
|
|
|
connection_t *conn;
|
|
|
|
|
|
|
|
for(i=0;i<nfds;i++) {
|
|
|
|
conn = connection_array[i];
|
2003-06-25 01:09:21 +02:00
|
|
|
if(conn->type == type && !conn->marked_for_close)
|
2003-06-25 09:19:30 +02:00
|
|
|
return conn;
|
2002-06-27 00:45:49 +02:00
|
|
|
}
|
major overhaul: dns slave subsystem, topics
on startup, it forks off a master dns handler, which forks off dns
slaves (like the apache model). slaves as spawned as load increases,
and then reused. excess slaves are not ever killed, currently.
implemented topics. each topic has a receive window in each direction
at each edge of the circuit, and sends sendme's at the data level, as
per before. each circuit also has receive windows in each direction at
each hop; an edge sends a circuit-level sendme as soon as enough data
cells have arrived (regardless of whether the data cells were flushed
to the exit conns). removed the 'connected' cell type, since it's now
a topic command within data cells.
at the edge of the circuit, there can be multiple connections associated
with a single circuit. you find them via the linked list conn->next_topic.
currently each new ap connection starts its own circuit, so we ought
to see comparable performance to what we had before. but that's only
because i haven't written the code to reattach to old circuits. please
try to break it as-is, and then i'll make it reuse the same circuit and
we'll try to break that.
svn:r152
2003-01-26 10:02:24 +01:00
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
2003-06-17 16:31:05 +02:00
|
|
|
connection_t *connection_get_by_type_state(int type, int state) {
|
|
|
|
int i;
|
|
|
|
connection_t *conn;
|
|
|
|
|
|
|
|
for(i=0;i<nfds;i++) {
|
|
|
|
conn = connection_array[i];
|
2003-06-25 01:09:21 +02:00
|
|
|
if(conn->type == type && conn->state == state && !conn->marked_for_close)
|
2003-06-25 09:19:30 +02:00
|
|
|
return conn;
|
2003-06-17 16:31:05 +02:00
|
|
|
}
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
2003-06-25 09:19:30 +02:00
|
|
|
connection_t *connection_get_by_type_state_lastwritten(int type, int state) {
|
|
|
|
int i;
|
|
|
|
connection_t *conn, *best=NULL;
|
|
|
|
|
|
|
|
for(i=0;i<nfds;i++) {
|
|
|
|
conn = connection_array[i];
|
|
|
|
if(conn->type == type && conn->state == state && !conn->marked_for_close)
|
|
|
|
if(!best || conn->timestamp_lastwritten < best->timestamp_lastwritten)
|
|
|
|
best = conn;
|
|
|
|
}
|
|
|
|
return best;
|
|
|
|
}
|
|
|
|
|
2002-06-27 00:45:49 +02:00
|
|
|
void connection_watch_events(connection_t *conn, short events) {
|
|
|
|
|
|
|
|
assert(conn && conn->poll_index < nfds);
|
|
|
|
|
|
|
|
poll_array[conn->poll_index].events = events;
|
|
|
|
}
|
|
|
|
|
2003-09-07 12:24:40 +02:00
|
|
|
int connection_is_reading(connection_t *conn) {
|
|
|
|
return poll_array[conn->poll_index].events & POLLIN;
|
|
|
|
}
|
|
|
|
|
Implemented link padding and receiver token buckets
Each socket reads at most 'bandwidth' bytes per second sustained, but
can handle bursts of up to 10*bandwidth bytes.
Cells are now sent out at evenly-spaced intervals, with padding sent
out otherwise. Set Linkpadding=0 in the rc file to send cells as soon
as they're available (and to never send padding cells).
Added license/copyrights statements at the top of most files.
router->min and router->max have been merged into a single 'bandwidth'
value. We should make the routerinfo_t reflect this (want to do that,
Mat?)
As the bandwidth increases, and we want to stop sleeping more and more
frequently to send a single cell, cpu usage goes up. At 128kB/s we're
pretty much calling poll with a timeout of 1ms or even 0ms. The current
code takes a timeout of 0-9ms and makes it 10ms. prepare_for_poll()
handles everything that should have happened in the past, so as long as
our buffers don't get too full in that 10ms, we're ok.
Speaking of too full, if you run three servers at 100kB/s with -l debug,
it spends too much time printing debugging messages to be able to keep
up with the cells. The outbuf ultimately fills up and it kills that
connection. If you run with -l err, it works fine up through 500kB/s and
probably beyond. Down the road we'll want to teach it to recognize when
an outbuf is getting full, and back off.
svn:r50
2002-07-16 03:12:15 +02:00
|
|
|
void connection_stop_reading(connection_t *conn) {
|
|
|
|
|
|
|
|
assert(conn && conn->poll_index < nfds);
|
|
|
|
|
2002-07-18 08:37:58 +02:00
|
|
|
log(LOG_DEBUG,"connection_stop_reading() called.");
|
Implemented link padding and receiver token buckets
Each socket reads at most 'bandwidth' bytes per second sustained, but
can handle bursts of up to 10*bandwidth bytes.
Cells are now sent out at evenly-spaced intervals, with padding sent
out otherwise. Set Linkpadding=0 in the rc file to send cells as soon
as they're available (and to never send padding cells).
Added license/copyrights statements at the top of most files.
router->min and router->max have been merged into a single 'bandwidth'
value. We should make the routerinfo_t reflect this (want to do that,
Mat?)
As the bandwidth increases, and we want to stop sleeping more and more
frequently to send a single cell, cpu usage goes up. At 128kB/s we're
pretty much calling poll with a timeout of 1ms or even 0ms. The current
code takes a timeout of 0-9ms and makes it 10ms. prepare_for_poll()
handles everything that should have happened in the past, so as long as
our buffers don't get too full in that 10ms, we're ok.
Speaking of too full, if you run three servers at 100kB/s with -l debug,
it spends too much time printing debugging messages to be able to keep
up with the cells. The outbuf ultimately fills up and it kills that
connection. If you run with -l err, it works fine up through 500kB/s and
probably beyond. Down the road we'll want to teach it to recognize when
an outbuf is getting full, and back off.
svn:r50
2002-07-16 03:12:15 +02:00
|
|
|
if(poll_array[conn->poll_index].events & POLLIN)
|
|
|
|
poll_array[conn->poll_index].events -= POLLIN;
|
|
|
|
}
|
|
|
|
|
|
|
|
void connection_start_reading(connection_t *conn) {
|
|
|
|
|
|
|
|
assert(conn && conn->poll_index < nfds);
|
|
|
|
|
|
|
|
poll_array[conn->poll_index].events |= POLLIN;
|
|
|
|
}
|
|
|
|
|
2002-07-18 08:37:58 +02:00
|
|
|
void connection_stop_writing(connection_t *conn) {
|
|
|
|
|
|
|
|
assert(conn && conn->poll_index < nfds);
|
|
|
|
|
|
|
|
if(poll_array[conn->poll_index].events & POLLOUT)
|
|
|
|
poll_array[conn->poll_index].events -= POLLOUT;
|
|
|
|
}
|
|
|
|
|
|
|
|
void connection_start_writing(connection_t *conn) {
|
|
|
|
|
|
|
|
assert(conn && conn->poll_index < nfds);
|
|
|
|
|
|
|
|
poll_array[conn->poll_index].events |= POLLOUT;
|
|
|
|
}
|
|
|
|
|
2003-07-05 07:46:06 +02:00
|
|
|
static void conn_read(int i) {
|
2003-09-28 08:48:20 +02:00
|
|
|
connection_t *conn = connection_array[i];
|
2002-06-27 00:45:49 +02:00
|
|
|
|
2003-09-28 08:48:20 +02:00
|
|
|
/* see http://www.greenend.org.uk/rjk/2001/06/poll.html for
|
|
|
|
* discussion of POLLIN vs POLLHUP */
|
2003-09-05 08:04:03 +02:00
|
|
|
if(!(poll_array[i].revents & (POLLIN|POLLHUP|POLLERR)))
|
2003-09-28 08:48:20 +02:00
|
|
|
if(!connection_speaks_cells(conn) ||
|
|
|
|
conn->state != OR_CONN_STATE_OPEN ||
|
|
|
|
!connection_is_reading(conn) ||
|
|
|
|
!tor_tls_get_pending_bytes(conn->tls))
|
|
|
|
return; /* this conn should not read */
|
2003-08-14 19:13:52 +02:00
|
|
|
|
cleanups, bugfixes, more verbose logs
Fixed up the assert_*_ok funcs some (more work remains)
Changed config so it reads either /etc/torrc or the -f arg, never both
Finally tracked down a nasty bug with our use of tls:
It turns out that if you ask SSL_read() for no more than n bytes, it
will read the entire record from the network (and maybe part of the next
record, I'm not sure), give you n bytes of it, and keep the remaining
bytes internally. This is fine, except our poll-for-read looks at the
network, and there are no bytes pending on the network, so we never know
to ask SSL_read() for more bytes. Currently I've hacked it so if we ask
for n bytes and it returns n bytes, then it reads again right then. This
will interact poorly with our rate limiting; we need a cleaner solution.
svn:r481
2003-09-24 23:24:52 +02:00
|
|
|
log_fn(LOG_DEBUG,"socket %d wants to read.",conn->s);
|
2003-09-23 21:47:41 +02:00
|
|
|
|
|
|
|
assert_connection_ok(conn, time(NULL));
|
Implemented link padding and receiver token buckets
Each socket reads at most 'bandwidth' bytes per second sustained, but
can handle bursts of up to 10*bandwidth bytes.
Cells are now sent out at evenly-spaced intervals, with padding sent
out otherwise. Set Linkpadding=0 in the rc file to send cells as soon
as they're available (and to never send padding cells).
Added license/copyrights statements at the top of most files.
router->min and router->max have been merged into a single 'bandwidth'
value. We should make the routerinfo_t reflect this (want to do that,
Mat?)
As the bandwidth increases, and we want to stop sleeping more and more
frequently to send a single cell, cpu usage goes up. At 128kB/s we're
pretty much calling poll with a timeout of 1ms or even 0ms. The current
code takes a timeout of 0-9ms and makes it 10ms. prepare_for_poll()
handles everything that should have happened in the past, so as long as
our buffers don't get too full in that 10ms, we're ok.
Speaking of too full, if you run three servers at 100kB/s with -l debug,
it spends too much time printing debugging messages to be able to keep
up with the cells. The outbuf ultimately fills up and it kills that
connection. If you run with -l err, it works fine up through 500kB/s and
probably beyond. Down the road we'll want to teach it to recognize when
an outbuf is getting full, and back off.
svn:r50
2002-07-16 03:12:15 +02:00
|
|
|
|
2003-09-05 08:04:03 +02:00
|
|
|
if(
|
|
|
|
/* XXX does POLLHUP also mean it's definitely broken? */
|
2003-08-14 19:13:52 +02:00
|
|
|
#ifdef MS_WINDOWS
|
2003-09-05 08:04:03 +02:00
|
|
|
(poll_array[i].revents & POLLERR) ||
|
2003-08-14 19:13:52 +02:00
|
|
|
#endif
|
2003-09-05 08:04:03 +02:00
|
|
|
connection_handle_read(conn) < 0)
|
|
|
|
{
|
|
|
|
/* this connection is broken. remove it */
|
|
|
|
log_fn(LOG_INFO,"%s connection broken, removing.", conn_type_to_string[conn->type]);
|
|
|
|
connection_remove(conn);
|
|
|
|
connection_free(conn);
|
|
|
|
if(i<nfds) { /* we just replaced the one at i with a new one. process it too. */
|
2003-07-05 07:46:06 +02:00
|
|
|
conn_read(i);
|
2003-09-05 08:04:03 +02:00
|
|
|
}
|
2003-09-23 21:47:41 +02:00
|
|
|
} else assert_connection_ok(conn, time(NULL));
|
2002-06-27 00:45:49 +02:00
|
|
|
}
|
|
|
|
|
2003-07-05 07:46:06 +02:00
|
|
|
static void conn_write(int i) {
|
2002-06-27 00:45:49 +02:00
|
|
|
connection_t *conn;
|
|
|
|
|
2003-09-05 08:04:03 +02:00
|
|
|
if(!(poll_array[i].revents & POLLOUT))
|
|
|
|
return; /* this conn doesn't want to write */
|
|
|
|
|
2003-07-05 07:46:06 +02:00
|
|
|
conn = connection_array[i];
|
cleanups, bugfixes, more verbose logs
Fixed up the assert_*_ok funcs some (more work remains)
Changed config so it reads either /etc/torrc or the -f arg, never both
Finally tracked down a nasty bug with our use of tls:
It turns out that if you ask SSL_read() for no more than n bytes, it
will read the entire record from the network (and maybe part of the next
record, I'm not sure), give you n bytes of it, and keep the remaining
bytes internally. This is fine, except our poll-for-read looks at the
network, and there are no bytes pending on the network, so we never know
to ask SSL_read() for more bytes. Currently I've hacked it so if we ask
for n bytes and it returns n bytes, then it reads again right then. This
will interact poorly with our rate limiting; we need a cleaner solution.
svn:r481
2003-09-24 23:24:52 +02:00
|
|
|
log_fn(LOG_DEBUG,"socket %d wants to write.",conn->s);
|
2002-06-27 00:45:49 +02:00
|
|
|
|
2003-09-23 21:47:41 +02:00
|
|
|
assert_connection_ok(conn, time(NULL));
|
|
|
|
|
2003-09-05 08:04:03 +02:00
|
|
|
if(connection_handle_write(conn) < 0) { /* this connection is broken. remove it. */
|
2003-09-26 12:03:50 +02:00
|
|
|
log_fn(LOG_INFO,"%s connection broken, removing.", conn_type_to_string[conn->type]);
|
2003-07-05 07:46:06 +02:00
|
|
|
connection_remove(conn);
|
|
|
|
connection_free(conn);
|
2003-09-05 08:04:03 +02:00
|
|
|
if(i<nfds) { /* we just replaced the one at i with a new one. process it too. */
|
2003-07-05 07:46:06 +02:00
|
|
|
conn_write(i);
|
2002-06-27 00:45:49 +02:00
|
|
|
}
|
2003-09-23 21:47:41 +02:00
|
|
|
} else assert_connection_ok(conn, time(NULL));
|
2002-06-27 00:45:49 +02:00
|
|
|
}
|
|
|
|
|
2003-07-05 07:46:06 +02:00
|
|
|
static void check_conn_marked(int i) {
|
2002-06-27 00:45:49 +02:00
|
|
|
connection_t *conn;
|
|
|
|
|
|
|
|
conn = connection_array[i];
|
2003-09-23 21:47:41 +02:00
|
|
|
assert_connection_ok(conn, time(NULL));
|
2002-06-27 00:45:49 +02:00
|
|
|
if(conn->marked_for_close) {
|
2003-09-26 12:03:50 +02:00
|
|
|
log_fn(LOG_INFO,"Cleaning up connection (fd %d).",conn->s);
|
2003-09-18 10:11:31 +02:00
|
|
|
if(conn->s >= 0) { /* might be an incomplete edge connection */
|
Integrated onion proxy into or/
The 'or' process can now be told (by the global_role variable) what
roles this server should play -- connect to all ORs, listen for ORs,
listen for OPs, listen for APs, or any combination.
* everything in /src/op/ is now obsolete.
* connection_ap.c now handles all interactions with application proxies
* "port" is now or_port, op_port, ap_port. But routers are still always
referenced (say, in conn_get_by_addr_port()) by addr / or_port. We
should make routers.c actually read these new ports (currently I've
kludged it so op_port = or_port+10, ap_port=or_port+20)
* circuits currently know if they're at the beginning of the path because
circ->cpath is set. They use this instead for crypts (both ways),
if it's set.
* I still obey the "send a 0 back to the AP when you're ready" protocol,
but I think we should phase it out. I can simply not read from the AP
socket until I'm ready.
I need to do a lot of cleanup work here, but the code appears to work, so
now's a good time for a checkin.
svn:r22
2002-07-02 11:36:58 +02:00
|
|
|
/* FIXME there's got to be a better way to check for this -- and make other checks? */
|
2003-09-25 12:42:07 +02:00
|
|
|
if(connection_speaks_cells(conn)) {
|
|
|
|
if(conn->state == OR_CONN_STATE_OPEN)
|
|
|
|
flush_buf_tls(conn->tls, conn->outbuf, &conn->outbuf_flushlen);
|
|
|
|
} else {
|
2003-09-25 07:17:11 +02:00
|
|
|
flush_buf(conn->s, conn->outbuf, &conn->outbuf_flushlen);
|
2003-09-25 12:42:07 +02:00
|
|
|
}
|
2003-09-18 10:11:31 +02:00
|
|
|
if(connection_wants_to_flush(conn)) /* not done flushing */
|
2003-09-25 07:17:11 +02:00
|
|
|
log_fn(LOG_WARNING,"Conn (socket %d) still wants to flush. Losing %d bytes!",conn->s, (int)buf_datalen(conn->inbuf));
|
Integrated onion proxy into or/
The 'or' process can now be told (by the global_role variable) what
roles this server should play -- connect to all ORs, listen for ORs,
listen for OPs, listen for APs, or any combination.
* everything in /src/op/ is now obsolete.
* connection_ap.c now handles all interactions with application proxies
* "port" is now or_port, op_port, ap_port. But routers are still always
referenced (say, in conn_get_by_addr_port()) by addr / or_port. We
should make routers.c actually read these new ports (currently I've
kludged it so op_port = or_port+10, ap_port=or_port+20)
* circuits currently know if they're at the beginning of the path because
circ->cpath is set. They use this instead for crypts (both ways),
if it's set.
* I still obey the "send a 0 back to the AP when you're ready" protocol,
but I think we should phase it out. I can simply not read from the AP
socket until I'm ready.
I need to do a lot of cleanup work here, but the code appears to work, so
now's a good time for a checkin.
svn:r22
2002-07-02 11:36:58 +02:00
|
|
|
}
|
2002-06-27 00:45:49 +02:00
|
|
|
connection_remove(conn);
|
|
|
|
connection_free(conn);
|
|
|
|
if(i<nfds) { /* we just replaced the one at i with a new one.
|
|
|
|
process it too. */
|
2002-07-05 08:27:23 +02:00
|
|
|
check_conn_marked(i);
|
2002-06-27 00:45:49 +02:00
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2003-09-05 08:04:03 +02:00
|
|
|
static int prepare_for_poll(void) {
|
Implemented link padding and receiver token buckets
Each socket reads at most 'bandwidth' bytes per second sustained, but
can handle bursts of up to 10*bandwidth bytes.
Cells are now sent out at evenly-spaced intervals, with padding sent
out otherwise. Set Linkpadding=0 in the rc file to send cells as soon
as they're available (and to never send padding cells).
Added license/copyrights statements at the top of most files.
router->min and router->max have been merged into a single 'bandwidth'
value. We should make the routerinfo_t reflect this (want to do that,
Mat?)
As the bandwidth increases, and we want to stop sleeping more and more
frequently to send a single cell, cpu usage goes up. At 128kB/s we're
pretty much calling poll with a timeout of 1ms or even 0ms. The current
code takes a timeout of 0-9ms and makes it 10ms. prepare_for_poll()
handles everything that should have happened in the past, so as long as
our buffers don't get too full in that 10ms, we're ok.
Speaking of too full, if you run three servers at 100kB/s with -l debug,
it spends too much time printing debugging messages to be able to keep
up with the cells. The outbuf ultimately fills up and it kills that
connection. If you run with -l err, it works fine up through 500kB/s and
probably beyond. Down the road we'll want to teach it to recognize when
an outbuf is getting full, and back off.
svn:r50
2002-07-16 03:12:15 +02:00
|
|
|
int i;
|
2003-09-28 08:48:20 +02:00
|
|
|
int timeout;
|
2003-09-07 12:24:40 +02:00
|
|
|
connection_t *conn;
|
2003-09-28 08:48:20 +02:00
|
|
|
struct timeval now;
|
2002-10-02 01:37:31 +02:00
|
|
|
static long current_second = 0; /* from previous calls to gettimeofday */
|
|
|
|
static long time_to_fetch_directory = 0;
|
2003-04-16 08:18:31 +02:00
|
|
|
static long time_to_new_circuit = 0;
|
2003-03-06 05:52:02 +01:00
|
|
|
// int ms_until_conn;
|
2002-10-02 01:37:31 +02:00
|
|
|
cell_t cell;
|
2003-04-16 08:18:31 +02:00
|
|
|
circuit_t *circ;
|
Implemented link padding and receiver token buckets
Each socket reads at most 'bandwidth' bytes per second sustained, but
can handle bursts of up to 10*bandwidth bytes.
Cells are now sent out at evenly-spaced intervals, with padding sent
out otherwise. Set Linkpadding=0 in the rc file to send cells as soon
as they're available (and to never send padding cells).
Added license/copyrights statements at the top of most files.
router->min and router->max have been merged into a single 'bandwidth'
value. We should make the routerinfo_t reflect this (want to do that,
Mat?)
As the bandwidth increases, and we want to stop sleeping more and more
frequently to send a single cell, cpu usage goes up. At 128kB/s we're
pretty much calling poll with a timeout of 1ms or even 0ms. The current
code takes a timeout of 0-9ms and makes it 10ms. prepare_for_poll()
handles everything that should have happened in the past, so as long as
our buffers don't get too full in that 10ms, we're ok.
Speaking of too full, if you run three servers at 100kB/s with -l debug,
it spends too much time printing debugging messages to be able to keep
up with the cells. The outbuf ultimately fills up and it kills that
connection. If you run with -l err, it works fine up through 500kB/s and
probably beyond. Down the road we'll want to teach it to recognize when
an outbuf is getting full, and back off.
svn:r50
2002-07-16 03:12:15 +02:00
|
|
|
|
2003-04-16 19:04:58 +02:00
|
|
|
my_gettimeofday(&now);
|
2003-09-28 08:48:20 +02:00
|
|
|
timeout = (1000 - (now.tv_usec / 1000)); /* how many milliseconds til the next second? */
|
Implemented link padding and receiver token buckets
Each socket reads at most 'bandwidth' bytes per second sustained, but
can handle bursts of up to 10*bandwidth bytes.
Cells are now sent out at evenly-spaced intervals, with padding sent
out otherwise. Set Linkpadding=0 in the rc file to send cells as soon
as they're available (and to never send padding cells).
Added license/copyrights statements at the top of most files.
router->min and router->max have been merged into a single 'bandwidth'
value. We should make the routerinfo_t reflect this (want to do that,
Mat?)
As the bandwidth increases, and we want to stop sleeping more and more
frequently to send a single cell, cpu usage goes up. At 128kB/s we're
pretty much calling poll with a timeout of 1ms or even 0ms. The current
code takes a timeout of 0-9ms and makes it 10ms. prepare_for_poll()
handles everything that should have happened in the past, so as long as
our buffers don't get too full in that 10ms, we're ok.
Speaking of too full, if you run three servers at 100kB/s with -l debug,
it spends too much time printing debugging messages to be able to keep
up with the cells. The outbuf ultimately fills up and it kills that
connection. If you run with -l err, it works fine up through 500kB/s and
probably beyond. Down the road we'll want to teach it to recognize when
an outbuf is getting full, and back off.
svn:r50
2002-07-16 03:12:15 +02:00
|
|
|
|
2003-03-06 05:52:02 +01:00
|
|
|
if(now.tv_sec > current_second) { /* the second has rolled over. check more stuff. */
|
|
|
|
|
2003-03-18 02:49:55 +01:00
|
|
|
if(!options.DirPort) {
|
2003-03-06 05:52:02 +01:00
|
|
|
if(time_to_fetch_directory < now.tv_sec) {
|
|
|
|
/* it's time to fetch a new directory */
|
|
|
|
/* NOTE directory servers do not currently fetch directories.
|
|
|
|
* Hope this doesn't bite us later.
|
|
|
|
*/
|
2003-09-23 21:47:41 +02:00
|
|
|
directory_initiate_command(router_pick_directory_server(),
|
|
|
|
DIR_CONN_STATE_CONNECTING_FETCH);
|
2003-03-06 05:52:02 +01:00
|
|
|
time_to_fetch_directory = now.tv_sec + options.DirFetchPeriod;
|
2002-09-26 14:09:10 +02:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2003-04-16 08:18:31 +02:00
|
|
|
if(options.APPort && time_to_new_circuit < now.tv_sec) {
|
|
|
|
circuit_expire_unused_circuits();
|
|
|
|
circuit_launch_new(-1); /* tell it to forget about previous failures */
|
2003-09-16 22:57:09 +02:00
|
|
|
circ = circuit_get_newest_open();
|
2003-04-16 08:18:31 +02:00
|
|
|
if(!circ || circ->dirty) {
|
2003-09-23 21:47:41 +02:00
|
|
|
log_fn(LOG_INFO,"Youngest circuit %s; launching replacement.", circ ? "dirty" : "missing");
|
2003-04-16 08:18:31 +02:00
|
|
|
circuit_launch_new(0); /* make an onion and lay the circuit */
|
|
|
|
}
|
|
|
|
time_to_new_circuit = now.tv_sec + options.NewCircuitPeriod;
|
|
|
|
}
|
|
|
|
|
2003-07-05 09:10:34 +02:00
|
|
|
if(global_read_bucket < 9*options.TotalBandwidth) {
|
|
|
|
global_read_bucket += options.TotalBandwidth;
|
|
|
|
log_fn(LOG_DEBUG,"global_read_bucket now %d.", global_read_bucket);
|
|
|
|
}
|
|
|
|
|
2003-03-06 05:52:02 +01:00
|
|
|
/* do housekeeping for each connection */
|
|
|
|
for(i=0;i<nfds;i++) {
|
2003-09-07 12:24:40 +02:00
|
|
|
conn = connection_array[i];
|
|
|
|
if(connection_receiver_bucket_should_increase(conn)) {
|
|
|
|
conn->receiver_bucket += conn->bandwidth;
|
|
|
|
// log_fn(LOG_DEBUG,"Receiver bucket %d now %d.", i, conn->receiver_bucket);
|
2003-07-05 09:10:34 +02:00
|
|
|
}
|
|
|
|
|
2003-09-07 12:24:40 +02:00
|
|
|
if(conn->wants_to_read == 1 /* it's marked to turn reading back on now */
|
2003-07-05 09:10:34 +02:00
|
|
|
&& global_read_bucket > 0 /* and we're allowed to read */
|
2003-09-27 23:09:56 +02:00
|
|
|
&& (!connection_speaks_cells(conn) || conn->receiver_bucket > 0)) {
|
|
|
|
/* and either a non-cell conn or a cell conn with non-empty bucket */
|
2003-09-07 12:24:40 +02:00
|
|
|
conn->wants_to_read = 0;
|
|
|
|
connection_start_reading(conn);
|
|
|
|
if(conn->wants_to_write == 1) {
|
|
|
|
conn->wants_to_write = 0;
|
|
|
|
connection_start_writing(conn);
|
|
|
|
}
|
2003-07-05 09:10:34 +02:00
|
|
|
}
|
2002-09-26 14:09:10 +02:00
|
|
|
|
2003-03-06 05:52:02 +01:00
|
|
|
/* check connections to see whether we should send a keepalive, expire, or wait */
|
2003-09-07 12:24:40 +02:00
|
|
|
if(!connection_speaks_cells(conn))
|
2003-03-06 05:52:02 +01:00
|
|
|
continue; /* this conn type doesn't send cells */
|
2003-09-28 08:48:20 +02:00
|
|
|
if(connection_state_is_open(conn) && tor_tls_get_pending_bytes(conn->tls))
|
|
|
|
timeout = 0; /* has pending bytes to read; don't let poll wait. */
|
2003-09-07 12:24:40 +02:00
|
|
|
if(now.tv_sec >= conn->timestamp_lastwritten + options.KeepalivePeriod) {
|
|
|
|
if((!options.OnionRouter && !circuit_get_by_conn(conn)) ||
|
|
|
|
(!connection_state_is_open(conn))) {
|
2003-03-06 05:52:02 +01:00
|
|
|
/* we're an onion proxy, with no circuits; or our handshake has expired. kill it. */
|
2003-09-26 12:03:50 +02:00
|
|
|
log_fn(LOG_INFO,"Expiring connection to %d (%s:%d).",
|
2003-09-07 12:24:40 +02:00
|
|
|
i,conn->address, conn->port);
|
|
|
|
conn->marked_for_close = 1;
|
2003-03-06 05:52:02 +01:00
|
|
|
} else {
|
|
|
|
/* either a full router, or we've got a circuit. send a padding cell. */
|
2003-09-23 21:47:41 +02:00
|
|
|
// log_fn(LOG_DEBUG,"Sending keepalive to (%s:%d)",
|
2003-09-07 12:24:40 +02:00
|
|
|
// conn->address, conn->port);
|
2003-03-11 22:38:38 +01:00
|
|
|
memset(&cell,0,sizeof(cell_t));
|
2003-03-06 05:52:02 +01:00
|
|
|
cell.command = CELL_PADDING;
|
2003-09-07 12:24:40 +02:00
|
|
|
if(connection_write_cell_to_buf(&cell, conn) < 0)
|
|
|
|
conn->marked_for_close = 1;
|
2003-03-06 05:52:02 +01:00
|
|
|
}
|
2002-10-02 01:37:31 +02:00
|
|
|
}
|
|
|
|
}
|
2003-03-06 05:52:02 +01:00
|
|
|
/* blow away any connections that need to die. can't do this later
|
2003-09-28 08:48:20 +02:00
|
|
|
* because we might open up a circuit and not realize we're about to cull
|
|
|
|
* the connection it's running over.
|
2003-03-06 05:52:02 +01:00
|
|
|
*/
|
|
|
|
for(i=0;i<nfds;i++)
|
|
|
|
check_conn_marked(i);
|
2002-10-02 01:37:31 +02:00
|
|
|
|
2002-12-31 16:04:14 +01:00
|
|
|
current_second = now.tv_sec; /* remember which second it is, for next time */
|
2002-10-02 01:37:31 +02:00
|
|
|
}
|
|
|
|
|
2003-09-28 08:48:20 +02:00
|
|
|
return timeout;
|
2003-03-06 05:52:02 +01:00
|
|
|
}
|
|
|
|
|
2003-09-25 07:17:11 +02:00
|
|
|
static crypto_pk_env_t *init_key_from_file(const char *fname)
|
|
|
|
{
|
|
|
|
crypto_pk_env_t *prkey = NULL;
|
|
|
|
int fd = -1;
|
|
|
|
FILE *file = NULL;
|
|
|
|
|
|
|
|
if (!(prkey = crypto_new_pk_env(CRYPTO_PK_RSA))) {
|
|
|
|
log(LOG_ERR, "Error creating crypto environment.");
|
|
|
|
goto error;
|
2002-06-27 00:45:49 +02:00
|
|
|
}
|
|
|
|
|
2003-09-26 20:27:35 +02:00
|
|
|
switch(file_status(fname)) {
|
2003-09-25 07:17:11 +02:00
|
|
|
case FN_DIR:
|
|
|
|
case FN_ERROR:
|
|
|
|
log(LOG_ERR, "Can't read key from %s", fname);
|
|
|
|
goto error;
|
|
|
|
case FN_NOENT:
|
|
|
|
log(LOG_INFO, "No key found in %s; generating fresh key.", fname);
|
|
|
|
if (crypto_pk_generate_key(prkey)) {
|
|
|
|
log(LOG_ERR, "Error generating key: %s", crypto_perror());
|
|
|
|
goto error;
|
2002-09-04 08:29:28 +02:00
|
|
|
}
|
2003-09-25 07:17:11 +02:00
|
|
|
if (crypto_pk_check_key(prkey) <= 0) {
|
|
|
|
log(LOG_ERR, "Generated key seems invalid");
|
|
|
|
goto error;
|
2003-05-08 00:40:03 +02:00
|
|
|
}
|
2003-09-25 07:17:11 +02:00
|
|
|
log(LOG_INFO, "Generated key seems valid");
|
2003-09-26 20:27:35 +02:00
|
|
|
if (crypto_pk_write_private_key_to_filename(prkey, fname)) {
|
|
|
|
log(LOG_ERR, "Couldn't write generated key to %s.", fname);
|
2003-09-25 07:17:11 +02:00
|
|
|
goto error;
|
|
|
|
}
|
|
|
|
return prkey;
|
|
|
|
case FN_FILE:
|
|
|
|
if (crypto_pk_read_private_key_from_filename(prkey, fname)) {
|
|
|
|
log(LOG_ERR, "Error loading private key.");
|
|
|
|
goto error;
|
2002-09-04 08:29:28 +02:00
|
|
|
}
|
2003-09-25 07:17:11 +02:00
|
|
|
return prkey;
|
|
|
|
default:
|
|
|
|
assert(0);
|
2002-06-27 00:45:49 +02:00
|
|
|
}
|
|
|
|
|
2003-09-25 07:17:11 +02:00
|
|
|
error:
|
|
|
|
if (prkey)
|
|
|
|
crypto_free_pk_env(prkey);
|
|
|
|
if (fd >= 0 && !file)
|
|
|
|
close(fd);
|
|
|
|
if (file)
|
|
|
|
fclose(file);
|
|
|
|
return NULL;
|
|
|
|
}
|
2003-09-08 08:26:38 +02:00
|
|
|
|
2003-09-25 07:17:11 +02:00
|
|
|
static int init_keys(void)
|
|
|
|
{
|
|
|
|
char keydir[512];
|
2003-09-26 20:27:35 +02:00
|
|
|
char fingerprint[FINGERPRINT_LEN+MAX_NICKNAME_LEN+3];
|
2003-09-25 07:17:11 +02:00
|
|
|
char *cp;
|
|
|
|
crypto_pk_env_t *prkey;
|
|
|
|
|
|
|
|
/* OP's don't need keys. Just initialize the TLS context.*/
|
|
|
|
if (!options.OnionRouter && !options.DirPort) {
|
|
|
|
if (tor_tls_context_new(NULL, 0, NULL)<0) {
|
|
|
|
log_fn(LOG_ERR, "Error creating TLS context for OP.");
|
2003-09-08 08:26:38 +02:00
|
|
|
return -1;
|
|
|
|
}
|
2003-09-25 07:17:11 +02:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
assert(options.DataDirectory);
|
|
|
|
if (strlen(options.DataDirectory) > (512-128)) {
|
|
|
|
log_fn(LOG_ERR, "DataDirectory is too long.");
|
|
|
|
return -1;
|
|
|
|
}
|
|
|
|
strcpy(keydir, options.DataDirectory);
|
2003-09-26 20:27:35 +02:00
|
|
|
if (check_private_dir(keydir, 1)) {
|
2003-09-25 07:17:11 +02:00
|
|
|
return -1;
|
|
|
|
}
|
|
|
|
strcat(keydir, "/keys");
|
2003-09-26 20:27:35 +02:00
|
|
|
if (check_private_dir(keydir, 1)) {
|
2003-09-25 07:17:11 +02:00
|
|
|
return -1;
|
|
|
|
}
|
|
|
|
cp = keydir + strlen(keydir); /* End of string. */
|
|
|
|
assert(!*cp);
|
|
|
|
|
|
|
|
/* 1. Read identity key. Make it if none is found. */
|
|
|
|
strcat(keydir, "/identity.key");
|
|
|
|
prkey = init_key_from_file(keydir);
|
|
|
|
if (!prkey) return -1;
|
|
|
|
set_identity_key(prkey);
|
|
|
|
/* 2. Read onion key. Make it if none is found. */
|
|
|
|
*cp = '\0';
|
|
|
|
strcat(keydir, "/onion.key");
|
|
|
|
prkey = init_key_from_file(keydir);
|
|
|
|
if (!prkey) return -1;
|
|
|
|
set_onion_key(prkey);
|
|
|
|
|
|
|
|
/* 3. Initialize link key and TLS context. */
|
|
|
|
*cp = '\0';
|
|
|
|
strcat(keydir, "/link.key");
|
|
|
|
prkey = init_key_from_file(keydir);
|
|
|
|
if (!prkey) return -1;
|
|
|
|
set_link_key(prkey);
|
|
|
|
if (tor_tls_context_new(prkey, 1, options.Nickname) < 0) {
|
|
|
|
log_fn(LOG_ERR, "Error initializing TLS context");
|
|
|
|
return -1;
|
|
|
|
}
|
|
|
|
/* 4. Dump router descriptor to 'router.desc' */
|
|
|
|
/* Must be called after keys are initialized. */
|
|
|
|
if (init_descriptor()<0) {
|
|
|
|
log_fn(LOG_ERR, "Error initializing descriptor.");
|
|
|
|
return -1;
|
|
|
|
}
|
|
|
|
strcpy(keydir, options.DataDirectory);
|
|
|
|
strcat(keydir, "/router.desc");
|
2003-09-26 20:27:35 +02:00
|
|
|
if (write_str_to_file(keydir, router_get_my_descriptor())) {
|
2003-09-25 07:17:11 +02:00
|
|
|
return -1;
|
|
|
|
}
|
|
|
|
/* 5. Dump fingerprint to 'fingerprint' */
|
|
|
|
strcpy(keydir, options.DataDirectory);
|
|
|
|
strcat(keydir, "/fingerprint");
|
2003-09-26 20:27:35 +02:00
|
|
|
assert(strlen(options.Nickname) <= MAX_NICKNAME_LEN);
|
|
|
|
strcpy(fingerprint, options.Nickname);
|
|
|
|
strcat(fingerprint, " ");
|
|
|
|
if (crypto_pk_get_fingerprint(get_identity_key(),
|
|
|
|
fingerprint+strlen(fingerprint))<0) {
|
2003-09-25 07:17:11 +02:00
|
|
|
log_fn(LOG_ERR, "Error computing fingerprint");
|
|
|
|
return -1;
|
|
|
|
}
|
2003-09-26 20:27:35 +02:00
|
|
|
strcat(fingerprint, "\n");
|
|
|
|
if (write_str_to_file(keydir, fingerprint))
|
|
|
|
return -1;
|
2003-09-25 07:17:11 +02:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int do_main_loop(void) {
|
|
|
|
int i;
|
|
|
|
int timeout;
|
|
|
|
int poll_result;
|
|
|
|
|
|
|
|
/* load the routers file */
|
|
|
|
if(router_get_list_from_file(options.RouterFile) < 0) {
|
|
|
|
log_fn(LOG_ERR,"Error loading router list.");
|
|
|
|
return -1;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* load the private keys, if we're supposed to have them, and set up the
|
|
|
|
* TLS context. */
|
|
|
|
if (init_keys() < 0) {
|
|
|
|
log_fn(LOG_ERR,"Error initializing keys; exiting");
|
|
|
|
return -1;
|
|
|
|
}
|
|
|
|
|
|
|
|
if(options.OnionRouter) {
|
2003-09-25 12:42:07 +02:00
|
|
|
cpu_init(); /* launch cpuworkers. Need to do this *after* we've read the onion key. */
|
2003-09-27 00:02:02 +02:00
|
|
|
router_upload_desc_to_dirservers(); /* upload our descriptor to all dirservers */
|
2003-09-08 08:26:38 +02:00
|
|
|
}
|
2003-09-07 12:24:40 +02:00
|
|
|
|
2003-03-18 02:49:55 +01:00
|
|
|
/* start up the necessary connections based on which ports are
|
|
|
|
* non-zero. This is where we try to connect to all the other ORs,
|
2003-09-05 08:04:03 +02:00
|
|
|
* and start the listeners.
|
2003-03-18 02:49:55 +01:00
|
|
|
*/
|
2003-08-12 17:08:51 +02:00
|
|
|
retry_all_connections((uint16_t) options.ORPort,
|
2003-08-21 01:05:22 +02:00
|
|
|
(uint16_t) options.APPort,
|
|
|
|
(uint16_t) options.DirPort);
|
2002-06-27 00:45:49 +02:00
|
|
|
|
|
|
|
for(;;) {
|
2003-08-12 17:08:51 +02:00
|
|
|
#ifndef MS_WIN32 /* do signal stuff only on unix */
|
2002-09-22 00:41:48 +02:00
|
|
|
if(please_dumpstats) {
|
|
|
|
dumpstats();
|
2002-09-28 03:40:11 +02:00
|
|
|
please_dumpstats = 0;
|
|
|
|
}
|
2003-09-21 08:15:43 +02:00
|
|
|
if(please_reset) {
|
|
|
|
/* fetch a new directory */
|
2003-03-18 02:49:55 +01:00
|
|
|
if(options.DirPort) {
|
2002-10-03 00:54:20 +02:00
|
|
|
if(router_get_list_from_file(options.RouterFile) < 0) {
|
2003-09-26 12:03:50 +02:00
|
|
|
log(LOG_WARNING,"Error reloading router list. Continuing with old list.");
|
2002-09-28 03:40:11 +02:00
|
|
|
}
|
|
|
|
} else {
|
2003-09-21 08:15:43 +02:00
|
|
|
directory_initiate_command(router_pick_directory_server(), DIR_CONN_STATE_CONNECTING_FETCH);
|
2002-09-28 03:40:11 +02:00
|
|
|
}
|
2003-09-21 08:15:43 +02:00
|
|
|
|
2003-09-23 21:47:41 +02:00
|
|
|
/* close and reopen the log files */
|
|
|
|
reset_logs();
|
2003-09-21 08:15:43 +02:00
|
|
|
|
|
|
|
please_reset = 0;
|
2002-09-22 00:41:48 +02:00
|
|
|
}
|
2003-08-12 08:41:53 +02:00
|
|
|
if(please_reap_children) {
|
|
|
|
while(waitpid(-1,NULL,WNOHANG)) ; /* keep reaping until no more zombies */
|
|
|
|
please_reap_children = 0;
|
|
|
|
}
|
2003-08-12 10:04:31 +02:00
|
|
|
#endif /* signal stuff */
|
Implemented link padding and receiver token buckets
Each socket reads at most 'bandwidth' bytes per second sustained, but
can handle bursts of up to 10*bandwidth bytes.
Cells are now sent out at evenly-spaced intervals, with padding sent
out otherwise. Set Linkpadding=0 in the rc file to send cells as soon
as they're available (and to never send padding cells).
Added license/copyrights statements at the top of most files.
router->min and router->max have been merged into a single 'bandwidth'
value. We should make the routerinfo_t reflect this (want to do that,
Mat?)
As the bandwidth increases, and we want to stop sleeping more and more
frequently to send a single cell, cpu usage goes up. At 128kB/s we're
pretty much calling poll with a timeout of 1ms or even 0ms. The current
code takes a timeout of 0-9ms and makes it 10ms. prepare_for_poll()
handles everything that should have happened in the past, so as long as
our buffers don't get too full in that 10ms, we're ok.
Speaking of too full, if you run three servers at 100kB/s with -l debug,
it spends too much time printing debugging messages to be able to keep
up with the cells. The outbuf ultimately fills up and it kills that
connection. If you run with -l err, it works fine up through 500kB/s and
probably beyond. Down the road we'll want to teach it to recognize when
an outbuf is getting full, and back off.
svn:r50
2002-07-16 03:12:15 +02:00
|
|
|
|
2003-09-05 08:04:03 +02:00
|
|
|
timeout = prepare_for_poll();
|
|
|
|
|
|
|
|
/* poll until we have an event, or the second ends */
|
Implemented link padding and receiver token buckets
Each socket reads at most 'bandwidth' bytes per second sustained, but
can handle bursts of up to 10*bandwidth bytes.
Cells are now sent out at evenly-spaced intervals, with padding sent
out otherwise. Set Linkpadding=0 in the rc file to send cells as soon
as they're available (and to never send padding cells).
Added license/copyrights statements at the top of most files.
router->min and router->max have been merged into a single 'bandwidth'
value. We should make the routerinfo_t reflect this (want to do that,
Mat?)
As the bandwidth increases, and we want to stop sleeping more and more
frequently to send a single cell, cpu usage goes up. At 128kB/s we're
pretty much calling poll with a timeout of 1ms or even 0ms. The current
code takes a timeout of 0-9ms and makes it 10ms. prepare_for_poll()
handles everything that should have happened in the past, so as long as
our buffers don't get too full in that 10ms, we're ok.
Speaking of too full, if you run three servers at 100kB/s with -l debug,
it spends too much time printing debugging messages to be able to keep
up with the cells. The outbuf ultimately fills up and it kills that
connection. If you run with -l err, it works fine up through 500kB/s and
probably beyond. Down the road we'll want to teach it to recognize when
an outbuf is getting full, and back off.
svn:r50
2002-07-16 03:12:15 +02:00
|
|
|
poll_result = poll(poll_array, nfds, timeout);
|
|
|
|
|
2002-07-16 20:24:12 +02:00
|
|
|
#if 0 /* let catch() handle things like ^c, and otherwise don't worry about it */
|
Implemented link padding and receiver token buckets
Each socket reads at most 'bandwidth' bytes per second sustained, but
can handle bursts of up to 10*bandwidth bytes.
Cells are now sent out at evenly-spaced intervals, with padding sent
out otherwise. Set Linkpadding=0 in the rc file to send cells as soon
as they're available (and to never send padding cells).
Added license/copyrights statements at the top of most files.
router->min and router->max have been merged into a single 'bandwidth'
value. We should make the routerinfo_t reflect this (want to do that,
Mat?)
As the bandwidth increases, and we want to stop sleeping more and more
frequently to send a single cell, cpu usage goes up. At 128kB/s we're
pretty much calling poll with a timeout of 1ms or even 0ms. The current
code takes a timeout of 0-9ms and makes it 10ms. prepare_for_poll()
handles everything that should have happened in the past, so as long as
our buffers don't get too full in that 10ms, we're ok.
Speaking of too full, if you run three servers at 100kB/s with -l debug,
it spends too much time printing debugging messages to be able to keep
up with the cells. The outbuf ultimately fills up and it kills that
connection. If you run with -l err, it works fine up through 500kB/s and
probably beyond. Down the road we'll want to teach it to recognize when
an outbuf is getting full, and back off.
svn:r50
2002-07-16 03:12:15 +02:00
|
|
|
if(poll_result < 0) {
|
|
|
|
log(LOG_ERR,"do_main_loop(): poll failed.");
|
|
|
|
if(errno != EINTR) /* let the program survive things like ^z */
|
|
|
|
return -1;
|
|
|
|
}
|
2002-07-16 20:24:12 +02:00
|
|
|
#endif
|
2002-06-27 00:45:49 +02:00
|
|
|
|
Implemented link padding and receiver token buckets
Each socket reads at most 'bandwidth' bytes per second sustained, but
can handle bursts of up to 10*bandwidth bytes.
Cells are now sent out at evenly-spaced intervals, with padding sent
out otherwise. Set Linkpadding=0 in the rc file to send cells as soon
as they're available (and to never send padding cells).
Added license/copyrights statements at the top of most files.
router->min and router->max have been merged into a single 'bandwidth'
value. We should make the routerinfo_t reflect this (want to do that,
Mat?)
As the bandwidth increases, and we want to stop sleeping more and more
frequently to send a single cell, cpu usage goes up. At 128kB/s we're
pretty much calling poll with a timeout of 1ms or even 0ms. The current
code takes a timeout of 0-9ms and makes it 10ms. prepare_for_poll()
handles everything that should have happened in the past, so as long as
our buffers don't get too full in that 10ms, we're ok.
Speaking of too full, if you run three servers at 100kB/s with -l debug,
it spends too much time printing debugging messages to be able to keep
up with the cells. The outbuf ultimately fills up and it kills that
connection. If you run with -l err, it works fine up through 500kB/s and
probably beyond. Down the road we'll want to teach it to recognize when
an outbuf is getting full, and back off.
svn:r50
2002-07-16 03:12:15 +02:00
|
|
|
if(poll_result > 0) { /* we have at least one connection to deal with */
|
2003-08-14 19:13:52 +02:00
|
|
|
/* do all the reads and errors first, so we can detect closed sockets */
|
Implemented link padding and receiver token buckets
Each socket reads at most 'bandwidth' bytes per second sustained, but
can handle bursts of up to 10*bandwidth bytes.
Cells are now sent out at evenly-spaced intervals, with padding sent
out otherwise. Set Linkpadding=0 in the rc file to send cells as soon
as they're available (and to never send padding cells).
Added license/copyrights statements at the top of most files.
router->min and router->max have been merged into a single 'bandwidth'
value. We should make the routerinfo_t reflect this (want to do that,
Mat?)
As the bandwidth increases, and we want to stop sleeping more and more
frequently to send a single cell, cpu usage goes up. At 128kB/s we're
pretty much calling poll with a timeout of 1ms or even 0ms. The current
code takes a timeout of 0-9ms and makes it 10ms. prepare_for_poll()
handles everything that should have happened in the past, so as long as
our buffers don't get too full in that 10ms, we're ok.
Speaking of too full, if you run three servers at 100kB/s with -l debug,
it spends too much time printing debugging messages to be able to keep
up with the cells. The outbuf ultimately fills up and it kills that
connection. If you run with -l err, it works fine up through 500kB/s and
probably beyond. Down the road we'll want to teach it to recognize when
an outbuf is getting full, and back off.
svn:r50
2002-07-16 03:12:15 +02:00
|
|
|
for(i=0;i<nfds;i++)
|
2003-09-05 08:04:03 +02:00
|
|
|
conn_read(i); /* this also blows away broken connections */
|
2002-06-27 00:45:49 +02:00
|
|
|
|
Implemented link padding and receiver token buckets
Each socket reads at most 'bandwidth' bytes per second sustained, but
can handle bursts of up to 10*bandwidth bytes.
Cells are now sent out at evenly-spaced intervals, with padding sent
out otherwise. Set Linkpadding=0 in the rc file to send cells as soon
as they're available (and to never send padding cells).
Added license/copyrights statements at the top of most files.
router->min and router->max have been merged into a single 'bandwidth'
value. We should make the routerinfo_t reflect this (want to do that,
Mat?)
As the bandwidth increases, and we want to stop sleeping more and more
frequently to send a single cell, cpu usage goes up. At 128kB/s we're
pretty much calling poll with a timeout of 1ms or even 0ms. The current
code takes a timeout of 0-9ms and makes it 10ms. prepare_for_poll()
handles everything that should have happened in the past, so as long as
our buffers don't get too full in that 10ms, we're ok.
Speaking of too full, if you run three servers at 100kB/s with -l debug,
it spends too much time printing debugging messages to be able to keep
up with the cells. The outbuf ultimately fills up and it kills that
connection. If you run with -l err, it works fine up through 500kB/s and
probably beyond. Down the road we'll want to teach it to recognize when
an outbuf is getting full, and back off.
svn:r50
2002-07-16 03:12:15 +02:00
|
|
|
/* then do the writes */
|
|
|
|
for(i=0;i<nfds;i++)
|
2003-09-05 08:04:03 +02:00
|
|
|
conn_write(i);
|
2002-06-27 00:45:49 +02:00
|
|
|
|
Implemented link padding and receiver token buckets
Each socket reads at most 'bandwidth' bytes per second sustained, but
can handle bursts of up to 10*bandwidth bytes.
Cells are now sent out at evenly-spaced intervals, with padding sent
out otherwise. Set Linkpadding=0 in the rc file to send cells as soon
as they're available (and to never send padding cells).
Added license/copyrights statements at the top of most files.
router->min and router->max have been merged into a single 'bandwidth'
value. We should make the routerinfo_t reflect this (want to do that,
Mat?)
As the bandwidth increases, and we want to stop sleeping more and more
frequently to send a single cell, cpu usage goes up. At 128kB/s we're
pretty much calling poll with a timeout of 1ms or even 0ms. The current
code takes a timeout of 0-9ms and makes it 10ms. prepare_for_poll()
handles everything that should have happened in the past, so as long as
our buffers don't get too full in that 10ms, we're ok.
Speaking of too full, if you run three servers at 100kB/s with -l debug,
it spends too much time printing debugging messages to be able to keep
up with the cells. The outbuf ultimately fills up and it kills that
connection. If you run with -l err, it works fine up through 500kB/s and
probably beyond. Down the road we'll want to teach it to recognize when
an outbuf is getting full, and back off.
svn:r50
2002-07-16 03:12:15 +02:00
|
|
|
/* any of the conns need to be closed now? */
|
|
|
|
for(i=0;i<nfds;i++)
|
|
|
|
check_conn_marked(i);
|
|
|
|
}
|
|
|
|
/* refilling buckets and sending cells happens at the beginning of the
|
|
|
|
* next iteration of the loop, inside prepare_for_poll()
|
|
|
|
*/
|
2002-06-27 00:45:49 +02:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2002-09-28 03:40:11 +02:00
|
|
|
static void catch(int the_signal) {
|
2002-06-27 00:45:49 +02:00
|
|
|
|
2003-08-12 17:08:51 +02:00
|
|
|
#ifndef MS_WIN32 /* do signal stuff only on unix */
|
2002-09-28 03:40:11 +02:00
|
|
|
switch(the_signal) {
|
2002-11-24 09:45:54 +01:00
|
|
|
// case SIGABRT:
|
2002-09-28 03:40:11 +02:00
|
|
|
case SIGTERM:
|
|
|
|
case SIGINT:
|
2003-09-26 12:03:50 +02:00
|
|
|
log(LOG_ERR,"Catching signal %d, exiting cleanly.", the_signal);
|
2002-09-28 03:40:11 +02:00
|
|
|
exit(0);
|
|
|
|
case SIGHUP:
|
2003-09-21 08:15:43 +02:00
|
|
|
please_reset = 1;
|
2002-09-28 03:40:11 +02:00
|
|
|
break;
|
|
|
|
case SIGUSR1:
|
|
|
|
please_dumpstats = 1;
|
|
|
|
break;
|
2003-08-12 08:41:53 +02:00
|
|
|
case SIGCHLD:
|
|
|
|
please_reap_children = 1;
|
2003-09-13 23:53:38 +02:00
|
|
|
break;
|
2002-09-28 03:40:11 +02:00
|
|
|
default:
|
2003-09-26 12:03:50 +02:00
|
|
|
log(LOG_WARNING,"Caught signal %d that we can't handle??", the_signal);
|
2002-09-28 03:40:11 +02:00
|
|
|
}
|
2003-08-12 10:04:31 +02:00
|
|
|
#endif /* signal stuff */
|
2002-09-22 00:41:48 +02:00
|
|
|
}
|
|
|
|
|
2003-07-05 07:46:06 +02:00
|
|
|
static void dumpstats(void) { /* dump stats to stdout */
|
2002-09-22 00:41:48 +02:00
|
|
|
int i;
|
|
|
|
connection_t *conn;
|
2002-10-02 01:37:31 +02:00
|
|
|
struct timeval now;
|
2002-09-22 00:41:48 +02:00
|
|
|
|
|
|
|
printf("Dumping stats:\n");
|
2003-04-16 19:04:58 +02:00
|
|
|
my_gettimeofday(&now);
|
2002-09-22 00:41:48 +02:00
|
|
|
|
|
|
|
for(i=0;i<nfds;i++) {
|
|
|
|
conn = connection_array[i];
|
2002-10-02 01:37:31 +02:00
|
|
|
printf("Conn %d (socket %d) type %d (%s), state %d (%s), created %ld secs ago\n",
|
2002-09-22 00:41:48 +02:00
|
|
|
i, conn->s, conn->type, conn_type_to_string[conn->type],
|
2002-10-02 01:37:31 +02:00
|
|
|
conn->state, conn_state_to_string[conn->type][conn->state], now.tv_sec - conn->timestamp_created);
|
2002-09-22 00:41:48 +02:00
|
|
|
if(!connection_is_listener(conn)) {
|
|
|
|
printf("Conn %d is to '%s:%d'.\n",i,conn->address, conn->port);
|
2003-09-25 07:17:11 +02:00
|
|
|
printf("Conn %d: %d bytes waiting on inbuf (last read %ld secs ago)\n",i,
|
|
|
|
(int)buf_datalen(conn->inbuf),
|
|
|
|
now.tv_sec - conn->timestamp_lastread);
|
|
|
|
printf("Conn %d: %d bytes waiting on outbuf (last written %ld secs ago)\n",i,(int)buf_datalen(conn->outbuf),
|
2002-10-02 01:37:31 +02:00
|
|
|
now.tv_sec - conn->timestamp_lastwritten);
|
2002-09-22 00:41:48 +02:00
|
|
|
}
|
|
|
|
circuit_dump_by_conn(conn); /* dump info about all the circuits using this conn */
|
|
|
|
printf("\n");
|
|
|
|
}
|
|
|
|
|
|
|
|
}
|
|
|
|
|
2003-09-25 07:17:11 +02:00
|
|
|
int dump_router_to_string(char *s, int maxlen, routerinfo_t *router,
|
|
|
|
crypto_pk_env_t *ident_key) {
|
|
|
|
char *onion_pkey;
|
|
|
|
char *link_pkey;
|
|
|
|
char *identity_pkey;
|
|
|
|
char digest[20];
|
|
|
|
char signature[128];
|
2003-09-26 20:27:35 +02:00
|
|
|
char published[32];
|
2003-09-25 07:17:11 +02:00
|
|
|
int onion_pkeylen, link_pkeylen, identity_pkeylen;
|
2002-09-26 14:09:10 +02:00
|
|
|
int written;
|
2003-04-08 08:44:38 +02:00
|
|
|
int result=0;
|
|
|
|
struct exit_policy_t *tmpe;
|
2002-09-26 15:17:14 +02:00
|
|
|
|
2003-09-25 07:17:11 +02:00
|
|
|
if(crypto_pk_write_public_key_to_string(router->onion_pkey,
|
|
|
|
&onion_pkey,&onion_pkeylen)<0) {
|
2003-09-26 12:03:50 +02:00
|
|
|
log_fn(LOG_WARNING,"write onion_pkey to string failed!");
|
2003-09-25 07:17:11 +02:00
|
|
|
return -1;
|
2002-09-26 15:17:14 +02:00
|
|
|
}
|
2003-04-08 08:44:38 +02:00
|
|
|
|
2003-09-25 07:17:11 +02:00
|
|
|
if(crypto_pk_write_public_key_to_string(router->identity_pkey,
|
|
|
|
&identity_pkey,&identity_pkeylen)<0) {
|
2003-09-26 12:03:50 +02:00
|
|
|
log_fn(LOG_WARNING,"write identity_pkey to string failed!");
|
2003-09-25 07:17:11 +02:00
|
|
|
return -1;
|
|
|
|
}
|
|
|
|
|
|
|
|
if(crypto_pk_write_public_key_to_string(router->link_pkey,
|
|
|
|
&link_pkey,&link_pkeylen)<0) {
|
2003-09-26 12:03:50 +02:00
|
|
|
log_fn(LOG_WARNING,"write link_pkey to string failed!");
|
2003-09-25 07:17:11 +02:00
|
|
|
return -1;
|
2003-05-07 04:13:23 +02:00
|
|
|
}
|
2003-09-26 20:27:35 +02:00
|
|
|
strftime(published, 32, "%Y-%m-%d %H:%M:%S", gmtime(&router->published_on));
|
2003-05-07 04:13:23 +02:00
|
|
|
|
2003-09-25 07:17:11 +02:00
|
|
|
result = snprintf(s, maxlen,
|
2003-09-27 23:30:10 +02:00
|
|
|
"router %s %s %d %d %d %d\n"
|
2003-09-26 20:27:35 +02:00
|
|
|
"published %s\n"
|
|
|
|
"onion-key\n%s"
|
2003-09-25 07:17:11 +02:00
|
|
|
"link-key\n%s"
|
|
|
|
"signing-key\n%s",
|
2003-09-27 23:30:10 +02:00
|
|
|
router->nickname,
|
2002-09-26 15:17:14 +02:00
|
|
|
router->address,
|
|
|
|
router->or_port,
|
|
|
|
router->ap_port,
|
|
|
|
router->dir_port,
|
|
|
|
router->bandwidth,
|
2003-09-26 20:27:35 +02:00
|
|
|
published,
|
2003-09-25 07:17:11 +02:00
|
|
|
onion_pkey, link_pkey, identity_pkey);
|
2002-09-26 15:17:14 +02:00
|
|
|
|
2003-09-25 07:17:11 +02:00
|
|
|
free(onion_pkey);
|
|
|
|
free(link_pkey);
|
|
|
|
free(identity_pkey);
|
2002-09-26 15:17:14 +02:00
|
|
|
|
2003-09-26 12:03:50 +02:00
|
|
|
if(result < 0 || result >= maxlen) {
|
2003-04-08 08:44:38 +02:00
|
|
|
/* apparently different glibcs do different things on snprintf error.. so check both */
|
|
|
|
return -1;
|
|
|
|
}
|
|
|
|
written = result;
|
|
|
|
|
|
|
|
for(tmpe=router->exit_policy; tmpe; tmpe=tmpe->next) {
|
|
|
|
result = snprintf(s+written, maxlen-written, "%s %s:%s\n",
|
|
|
|
tmpe->policy_type == EXIT_POLICY_ACCEPT ? "accept" : "reject",
|
|
|
|
tmpe->address, tmpe->port);
|
|
|
|
if(result < 0 || result+written > maxlen) {
|
|
|
|
/* apparently different glibcs do different things on snprintf error.. so check both */
|
|
|
|
return -1;
|
|
|
|
}
|
|
|
|
written += result;
|
|
|
|
}
|
2003-09-25 07:17:11 +02:00
|
|
|
if (written > maxlen-256) /* Not enough room for signature. */
|
|
|
|
return -1;
|
2003-04-08 08:44:38 +02:00
|
|
|
|
2003-09-25 07:17:11 +02:00
|
|
|
strcat(s+written, "router-signature\n");
|
|
|
|
written += strlen(s+written);
|
|
|
|
s[written] = '\0';
|
|
|
|
if (router_get_router_hash(s, digest) < 0)
|
|
|
|
return -1;
|
2003-04-08 08:44:38 +02:00
|
|
|
|
2003-09-25 07:17:11 +02:00
|
|
|
if (crypto_pk_private_sign(ident_key, digest, 20, signature) < 0) {
|
2003-09-26 12:03:50 +02:00
|
|
|
log_fn(LOG_WARNING, "Error signing digest");
|
2003-09-25 07:17:11 +02:00
|
|
|
return -1;
|
|
|
|
}
|
|
|
|
strcat(s+written, "-----BEGIN SIGNATURE-----\n");
|
|
|
|
written += strlen(s+written);
|
|
|
|
if (base64_encode(s+written, maxlen-written, signature, 128) < 0) {
|
2003-09-26 12:03:50 +02:00
|
|
|
log_fn(LOG_WARNING, "Couldn't base64-encode signature");
|
|
|
|
/* XXX Nick: do we really mean to fall through here? */
|
2003-09-25 07:17:11 +02:00
|
|
|
}
|
|
|
|
written += strlen(s+written);
|
|
|
|
strcat(s+written, "-----END SIGNATURE-----\n");
|
|
|
|
written += strlen(s+written);
|
|
|
|
|
|
|
|
if (written > maxlen-2)
|
|
|
|
return -1;
|
2003-04-08 08:44:38 +02:00
|
|
|
/* include a last '\n' */
|
|
|
|
s[written] = '\n';
|
|
|
|
s[written+1] = 0;
|
|
|
|
return written+1;
|
2002-09-26 15:17:14 +02:00
|
|
|
}
|
|
|
|
|
2003-09-27 23:30:10 +02:00
|
|
|
int
|
|
|
|
list_running_servers(char **nicknames_out)
|
|
|
|
{
|
|
|
|
char *nickname_lst[MAX_ROUTERS_IN_DIR];
|
2002-09-26 15:17:14 +02:00
|
|
|
connection_t *conn;
|
2003-09-27 23:30:10 +02:00
|
|
|
char *cp;
|
|
|
|
int n = 0, i;
|
|
|
|
int length;
|
|
|
|
*nicknames_out = NULL;
|
|
|
|
if (my_routerinfo)
|
|
|
|
nickname_lst[n++] = my_routerinfo->nickname;
|
|
|
|
for (i = 0; i<nfds; ++i) {
|
2002-09-26 14:09:10 +02:00
|
|
|
conn = connection_array[i];
|
2003-09-27 23:30:10 +02:00
|
|
|
if (conn->type != CONN_TYPE_OR || conn->state != OR_CONN_STATE_OPEN)
|
|
|
|
continue; /* only list successfully handshaked OR's. */
|
|
|
|
nickname_lst[n++] = conn->nickname;
|
|
|
|
}
|
|
|
|
length = n + 1; /* spaces + EOS + 1. */
|
|
|
|
for (i = 0; i<n; ++i) {
|
|
|
|
length += strlen(nickname_lst[i]);
|
|
|
|
}
|
|
|
|
*nicknames_out = tor_malloc(length);
|
2003-09-28 08:48:20 +02:00
|
|
|
log_fn(LOG_DEBUG,"total length %d malloced.",length);
|
2003-09-27 23:30:10 +02:00
|
|
|
cp = *nicknames_out;
|
|
|
|
for (i = 0; i<n; ++i) {
|
|
|
|
if (i)
|
|
|
|
strcat(cp, " ");
|
|
|
|
strcat(cp, nickname_lst[i]);
|
|
|
|
while (*cp)
|
|
|
|
++cp;
|
2003-09-28 08:48:20 +02:00
|
|
|
log_fn(LOG_DEBUG,"end of loop %d, now %d written (nick %s)",
|
|
|
|
i,1+(int)(cp-*nicknames_out),nickname_lst[i]);
|
Get directories working.
Or at least, directories get generated, signed, download, and checked, with
nobody seeming to crash.
In config/*, added 'signing-key' blocks to dirservers and routers.or, so
that everyone will know about the directories' signing keys.
In or/directory.c, refrained from using a dirserver's signing key when
no such key is known; added more debugging output.
In or/main.c, added debugging output and fixed a few logic errors.
In or/routers.c, added debugging output and prevented a segfault on
routers_resolve_directory. The interleaving of arrays and lists on
routerinfo_t is still messy, but at least it seems to work again.
svn:r278
2003-05-08 23:35:11 +02:00
|
|
|
}
|
2003-05-07 04:13:23 +02:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2003-09-25 07:17:11 +02:00
|
|
|
static char descriptor[8192];
|
|
|
|
/* XXX should this replace my_routerinfo? */
|
|
|
|
static routerinfo_t *desc_routerinfo;
|
|
|
|
const char *router_get_my_descriptor(void) {
|
2003-09-25 12:42:07 +02:00
|
|
|
log_fn(LOG_DEBUG,"my desc is '%s'",descriptor);
|
2003-09-25 07:17:11 +02:00
|
|
|
return descriptor;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int init_descriptor(void) {
|
|
|
|
routerinfo_t *ri;
|
2003-09-25 12:42:07 +02:00
|
|
|
char localhostname[256];
|
2003-09-27 09:21:36 +02:00
|
|
|
char *address = options.Address;
|
2003-09-25 12:42:07 +02:00
|
|
|
|
2003-09-27 09:21:36 +02:00
|
|
|
if(!address) { /* if not specified in config, we find a default */
|
|
|
|
if(gethostname(localhostname,sizeof(localhostname)) < 0) {
|
|
|
|
log_fn(LOG_WARNING,"Error obtaining local hostname");
|
|
|
|
return -1;
|
|
|
|
}
|
|
|
|
address = localhostname;
|
2003-09-25 12:42:07 +02:00
|
|
|
}
|
2003-09-25 07:17:11 +02:00
|
|
|
ri = tor_malloc(sizeof(routerinfo_t));
|
2003-09-27 09:21:36 +02:00
|
|
|
ri->address = strdup(address);
|
2003-09-25 07:17:11 +02:00
|
|
|
ri->nickname = strdup(options.Nickname);
|
2003-09-27 09:21:36 +02:00
|
|
|
/* No need to set addr. */
|
2003-09-25 07:17:11 +02:00
|
|
|
ri->or_port = options.ORPort;
|
|
|
|
ri->ap_port = options.APPort;
|
|
|
|
ri->dir_port = options.DirPort;
|
2003-09-26 20:27:35 +02:00
|
|
|
ri->published_on = time(NULL);
|
2003-09-25 07:17:11 +02:00
|
|
|
ri->onion_pkey = crypto_pk_dup_key(get_onion_key());
|
|
|
|
ri->link_pkey = crypto_pk_dup_key(get_link_key());
|
|
|
|
ri->identity_pkey = crypto_pk_dup_key(get_identity_key());
|
|
|
|
ri->bandwidth = options.TotalBandwidth;
|
|
|
|
ri->exit_policy = NULL; /* XXX implement this. */
|
|
|
|
if (desc_routerinfo)
|
|
|
|
routerinfo_free(desc_routerinfo);
|
|
|
|
desc_routerinfo = ri;
|
|
|
|
if (dump_router_to_string(descriptor, 8192, ri, get_identity_key())<0) {
|
2003-09-26 12:03:50 +02:00
|
|
|
log_fn(LOG_WARNING, "Couldn't dump router to string.");
|
2003-09-25 07:17:11 +02:00
|
|
|
return -1;
|
|
|
|
}
|
|
|
|
return 0;
|
2003-09-17 22:09:06 +02:00
|
|
|
}
|
|
|
|
|
2003-04-16 08:18:31 +02:00
|
|
|
void daemonize(void) {
|
2003-08-12 17:08:51 +02:00
|
|
|
#ifndef MS_WINDOWS
|
2003-03-17 03:41:36 +01:00
|
|
|
/* Fork; parent exits. */
|
|
|
|
if (fork())
|
|
|
|
exit(0);
|
|
|
|
|
|
|
|
/* Create new session; make sure we never get a terminal */
|
|
|
|
setsid();
|
|
|
|
if (fork())
|
|
|
|
exit(0);
|
|
|
|
|
|
|
|
chdir("/");
|
|
|
|
umask(000);
|
|
|
|
|
|
|
|
fclose(stdin);
|
2003-06-18 07:52:32 +02:00
|
|
|
fclose(stdout); /* XXX Nick: this closes our log, right? is it safe to leave this open? */
|
2003-03-17 03:41:36 +01:00
|
|
|
fclose(stderr);
|
2003-08-12 17:08:51 +02:00
|
|
|
#endif
|
2003-03-17 03:41:36 +01:00
|
|
|
}
|
|
|
|
|
2003-04-07 04:12:02 +02:00
|
|
|
int tor_main(int argc, char *argv[]) {
|
2002-06-27 00:45:49 +02:00
|
|
|
int retval = 0;
|
|
|
|
|
2002-11-23 07:49:01 +01:00
|
|
|
if(getconfig(argc,argv,&options))
|
|
|
|
exit(1);
|
2003-06-21 21:03:22 +02:00
|
|
|
log_set_severity(options.loglevel); /* assign logging severity level from options */
|
2003-07-05 09:10:34 +02:00
|
|
|
global_read_bucket = options.TotalBandwidth; /* start it at 1 second of traffic */
|
2002-06-27 00:45:49 +02:00
|
|
|
|
2003-06-17 16:31:05 +02:00
|
|
|
if(options.Daemon)
|
2003-03-17 03:41:36 +01:00
|
|
|
daemonize();
|
|
|
|
|
2003-05-28 04:03:25 +02:00
|
|
|
if(options.OnionRouter) { /* only spawn dns handlers if we're a router */
|
2003-06-17 16:31:05 +02:00
|
|
|
dns_init(); /* initialize the dns resolve tree, and spawn workers */
|
major overhaul: dns slave subsystem, topics
on startup, it forks off a master dns handler, which forks off dns
slaves (like the apache model). slaves as spawned as load increases,
and then reused. excess slaves are not ever killed, currently.
implemented topics. each topic has a receive window in each direction
at each edge of the circuit, and sends sendme's at the data level, as
per before. each circuit also has receive windows in each direction at
each hop; an edge sends a circuit-level sendme as soon as enough data
cells have arrived (regardless of whether the data cells were flushed
to the exit conns). removed the 'connected' cell type, since it's now
a topic command within data cells.
at the edge of the circuit, there can be multiple connections associated
with a single circuit. you find them via the linked list conn->next_topic.
currently each new ap connection starts its own circuit, so we ought
to see comparable performance to what we had before. but that's only
because i haven't written the code to reattach to old circuits. please
try to break it as-is, and then i'll make it reuse the same circuit and
we'll try to break that.
svn:r152
2003-01-26 10:02:24 +01:00
|
|
|
}
|
|
|
|
|
2003-08-12 10:04:31 +02:00
|
|
|
#ifndef MS_WINDOWS /* do signal stuff only on unix */
|
major overhaul: dns slave subsystem, topics
on startup, it forks off a master dns handler, which forks off dns
slaves (like the apache model). slaves as spawned as load increases,
and then reused. excess slaves are not ever killed, currently.
implemented topics. each topic has a receive window in each direction
at each edge of the circuit, and sends sendme's at the data level, as
per before. each circuit also has receive windows in each direction at
each hop; an edge sends a circuit-level sendme as soon as enough data
cells have arrived (regardless of whether the data cells were flushed
to the exit conns). removed the 'connected' cell type, since it's now
a topic command within data cells.
at the edge of the circuit, there can be multiple connections associated
with a single circuit. you find them via the linked list conn->next_topic.
currently each new ap connection starts its own circuit, so we ought
to see comparable performance to what we had before. but that's only
because i haven't written the code to reattach to old circuits. please
try to break it as-is, and then i'll make it reuse the same circuit and
we'll try to break that.
svn:r152
2003-01-26 10:02:24 +01:00
|
|
|
signal (SIGINT, catch); /* catch kills so we can exit cleanly */
|
|
|
|
signal (SIGTERM, catch);
|
|
|
|
signal (SIGUSR1, catch); /* to dump stats to stdout */
|
|
|
|
signal (SIGHUP, catch); /* to reload directory */
|
2003-08-12 08:41:53 +02:00
|
|
|
signal (SIGCHLD, catch); /* for exiting dns/cpu workers */
|
2003-08-12 10:04:31 +02:00
|
|
|
#endif /* signal stuff */
|
major overhaul: dns slave subsystem, topics
on startup, it forks off a master dns handler, which forks off dns
slaves (like the apache model). slaves as spawned as load increases,
and then reused. excess slaves are not ever killed, currently.
implemented topics. each topic has a receive window in each direction
at each edge of the circuit, and sends sendme's at the data level, as
per before. each circuit also has receive windows in each direction at
each hop; an edge sends a circuit-level sendme as soon as enough data
cells have arrived (regardless of whether the data cells were flushed
to the exit conns). removed the 'connected' cell type, since it's now
a topic command within data cells.
at the edge of the circuit, there can be multiple connections associated
with a single circuit. you find them via the linked list conn->next_topic.
currently each new ap connection starts its own circuit, so we ought
to see comparable performance to what we had before. but that's only
because i haven't written the code to reattach to old circuits. please
try to break it as-is, and then i'll make it reuse the same circuit and
we'll try to break that.
svn:r152
2003-01-26 10:02:24 +01:00
|
|
|
|
|
|
|
crypto_global_init();
|
2003-06-13 23:13:37 +02:00
|
|
|
crypto_seed_rng();
|
2002-06-27 00:45:49 +02:00
|
|
|
retval = do_main_loop();
|
2002-08-22 09:30:03 +02:00
|
|
|
crypto_global_cleanup();
|
2003-09-25 07:17:11 +02:00
|
|
|
|
2002-06-27 00:45:49 +02:00
|
|
|
return retval;
|
|
|
|
}
|
|
|
|
|
2003-04-07 04:12:02 +02:00
|
|
|
/*
|
|
|
|
Local Variables:
|
|
|
|
mode:c
|
|
|
|
indent-tabs-mode:nil
|
|
|
|
c-basic-offset:2
|
|
|
|
End:
|
|
|
|
*/
|